Module 5
Module 5
Module 5
Also called “top-down” information, hence is at the top of the figure with an
arrow pointing downwards
So, “th” (/ð/) of “the”, “e” (/ə/), the “l” (/l/) of “librarian”, “i” (/aɪ/), “b” (/b/) and
so on.
Then, the speech variables not included in phonetic segments, also known as
suprasegmentals, come under prosodic analysis.
As the word is uttered, it activates the listener’s lexicon, which includes all
words that match the speech signal.
As the information compiles little by little, the mental lexicon removes the
unlikely candidates, and narrows down to the final possibilities among which it
selects the word in question.
Once the word is mapped in the lexicon, information regarding the meaning,
the morphology, the grammatical category, and the syntactic and semantic
structures occurs in the listener.
Syntactic and semantic processing:
Here, the syntactic structure of the sentence is determined. This process is
known as parsing.
So,
Next comes the assignment of the thematic roles. This involves assigning the
semantic roles to the phrases in relation to the verb.
Though the processes are mentioned in a serial manner here, they actually occur
quasi-simultaneously. These are near parallel processes. (Marslen-Wilson,1975)
These are online processes, i.e. they occur as the utterance is taking place.
It is also generally agreed that these layers of different processes are
interactive, i.e. there is an exchange of information between the levels of
processes.
These basic processing components and levels are found in both monolingual
and bilingual speech comprehension.
Thus, the phoneme inventories may not overlap entirely, word order
may be different, morphological processes differ and so on.
Second, in most cases the knowledge in the languages might differ based on
proficiency. Hence that will impact their comprehension and perception.
For example, the case of tones. If a bilingual speaks tow languages, one of
which is tonal, the strategies will be different in each language. Lexical
processing in one will need to take tons into account but not in the other.
Similarly, differences at other levels, like syntax, will also affect the relative
processing speed and strategy.
Fourth,
Bilingual mode will be activated when it would depend on the specific requirements
of the system and the user's needs.
For example, if the user communicates in two languages, bilingual mode may need
to be activated to allow for accurate understanding and response in both languages.
Eye tracking study: participants were asked to click on the picture that
matched the auditory stimuli
Manipulation: changing the Voice Onset Time (VOT) of the consonant ‘p’, to
make it sound like English ‘p’.
This study tried to create a monolingual mode by have experimenters who spoke
either Russian or English (posing as monolinguals)
Also other changes were made to the contextual factors, like conducting Russian
and English experiments in different sessions etc.
For example, if the dominant language has only one category corresponding to two of
the weaker one, then the two might get merged.
Thus English / æ/ and / / get assimilated into Dutch / /
In the bilingual mode: both languages are active, but base language network is more
strongly activated.
And, if an input unit has similarities in the other language, this will lead to delayed
processing.
BIMOLA (Lewy & Grosjean 2008)
Bilingual Interactive Model of Lexical
Access:
BIA is based on Dutch and English, whereas BIMOLA simulates French and English
words.
There are some other important internal feature dependent differences too.
However, later studies showed the key role bilinguals played to understand
the effects of language experience on speech perception.
For example:
Simultaneous bilinguals: exposed to both languages within first year of life
Heritage speakers: AOA corresponds to age when they first attend preschool
or school.
Immigrant population: AOA used alternatively with “age of arrival at host
country”. Might occur in childhood, or adulthood.
Flege, MacKay and Meador (1999)
Examined perception of English vowels by Italian-English bilinguals of varying
AOA in Canada.
72 Participants of varying AOA:
Early (7 years old)
Mid (14 years old)
Late (19 years old)
L1:L2 usage ratio refers to respective amount of communication that takes place
in an individual’s L1 and L2
High levels of continued L1 usage might reflect limited exposure to L2 and vice
versa.
Tasks included:
Task 1, 2 & 3: discriminate several English vowel contrast
Task 4: identify vowels in phrases they heard and read and indicate if they were
produced correctly. These contrasts typically did not occur in Italian.
Results:
“Bilinguals most often have a dominant, or stronger, language in which they have
attained an overall higher level of proficiency, which they use more frequently, and
across a wider range of domains.” [Silva-Corvalán and Treffers-Daller, 2016]
Language dominance of bilinguals is not fixed but can change over time.
As bilingual’s environment changes, so will their need for that particular language skills.
Amengual and Chamorro (2015)
Aim: investigate role of language dominance and see if the Galician vowel
system becomes more Spanish-like due to extended exposure to Spanish.
Participants:
English monolinguals
Spanish-English late bilinguals
Early Spanish-English bilinguals
Language mode (Grosjean 2001, 2008) refers to the state of activation of the
languages of a bilingual, and the language processing mechanisms at a given point in
time.
Task: Participants had to identify the deviant stimuli from repeated sequence of
standard stimuli. MMN was used to observe the response.
Young infants are called “universal listener” because of their broad speech
perception, which helps them acquire any native language.
Over time, listeners start to specialize in attending to just the sound contrasts of
their own languages, called perceptual narrowing.
Have less exposure to each language than monolinguals since their time
is divided between the 2 languages
Source: https://static1.squarespace.com/static/555906d9e4b0251d92d5635a/t/
5bb7d0c7104c7b66055d47b8/1538773192178/2018_Byers-Heinlein_Chapter8_SpeechPerception.pdf
Kuipers and Thierry (2012)
Investigated if the English-Welsh bilinguals could discriminate between
words in each of the languages.
Task:
Children shown series of familiar pictures labeled in one or the other language.
Several trials of one language followed by one trial in other language.
Result:
ERP showed that the bilinguals and the control group of English monolingual
toddlers detected language change
However, response timings showed a difference, suggesting difference in
processing.
Gervain and Werker (2013)
Source: https://static1.squarespace.com/static/555906d9e4b0251d92d5635a/t/
5bb7d0c7104c7b66055d47b8/1538773192178/2018_Byers-Heinlein_Chapter8_SpeechPerception.pdf
Task:
Infants first made to hear stream of nonsense syllables consisting of
frequent and infrequent syllables.
Then, they were played syllable stream that were characteristic of either
OV or VO parsing strategy.
Results:
When bilinguals heard stream with OV prosody; they tend to look longer or surprised.
When they heard stream with VO prosody, they showed the opposite pattern of
looking.
Meanwhile, the monolinguals were not able to parse the stream of OV prosody, and
defaulted to VO parsing strategy, even when they heard stream with no prosody.
These results together suggest that bilingual infants separate their languages
based on prosody, and use this information to form expectations about the
word order they are hearing.
Consonant Perception
By the end of their first year, bilingual infants perceive many consonant
contrasts that are important in their native languages.
Source: https://static1.squarespace.com/static/555906d9e4b0251d92d5635a/t/
5bb7d0c7104c7b66055d47b8/1538773192178/2018_Byers-Heinlein_Chapter8_SpeechPerception.pdf
Measurement was done in two time windows: 100-260 ms & 260-460 ms
after the stimulus.
Results:
For /t/ (the English contrast), no differences found between monolinguals and
bilinguals in either time window.
for /d/ (the Spanish contrast), bilinguals showed a stronger neural response
than the monolinguals in both time windows.
Task:
Infants watched an animated Elmo character disappear into a t‐shaped
tunnel, and then emerge on one side or the other.
The /e/ sound predicted Elmo’s reappearance on the right side, while hearing
the /ε/ sound predicted Elmo’s reappearance on the left side (or vice versa).
Result: Both monolingual Catalan and bilingual Catalan–Spanish infants
succeeded, which they could only do if they perceived the difference
between the two sounds.
Child L2 Learners
Studies on phonological development in child L2 learners predate similar
studies on bilingual infants as:
Children can follow instructions, push buttons, and answer verbally.
Whereas infants’ phonological knowledge must be inferred indirectly from
brain responses and looking time.
Results:
All groups improved over the course of testing.
Children aged 12–15 did the best of any group, and performed similarly to
native Dutch‐speaking children, even during the first testing session – a short
6 months after arriving in the Netherlands.
Task:
Children heard a word, across several trials, and had to point to a picture
representing the word they had heard.
Children were presented with words across a range of voice onset time
values, so that researchers could pinpoint where they placed their boundaries.
Results:
The native English speakers and the L2 learners showed very similar
performances overall.
These results would suggest that children quickly acquire the perceptual skills
to process L2 consonants.
Aoyama et al. 2008
This study compared Japanese children and adults learning of English
consonant sounds.
The consonants investigated were the sounds /s/ (as in “sink”) and /θ/ (as
in “think”).
Task:
The participants heard three sounds and had to indicate which of the three was
different from the other two.
Both groups were tested twice: first, around six months after arrival in the
United States and then around 1.5 years after arrival.
Results:
Both the Japanese children and adults showed difficulty with this task
compared to native English‐speaking adults and children.
These results suggest that it may take some time for child L2 learners to
accurately perceive some L2 speech sounds.
Vowel perception
Participants: Children aged 9–12 years and had started learning German
in preschool.
Task:
Children tested on their discrimination of several different German vowels
previously classified as either difficult or easy for monolingual Turkish adults to
discriminate.
Children heard three different robots each produce a nonsense word (e.g.,
“kak”, “kak”, “kek”), and had to decide which robot was saying something
different.
Results:
Bilingual children were less able to discriminate the difficult contrasts, but
equally able to discriminate the easy contrasts, as compared to the control
group of German monolingual children.
Suggesting that despite their early and ongoing exposure to German over
many years, they still perceived speech sounds differently than monolinguals.
Results:
Bilingual children perceived the vowels more accurately than bilingual adults,
but were less accurate than monolingual English‐speaking children.
Children who had been in the country for a longer duration did significantly
better than those who had been in the country for a shorter duration, while
this had no effect for adults.
The study also revealed that unlike their performance on the perception
task, bilingual children’s productions of the same vowels were just as
accurate as monolingual children’s productions.
Base language effect in categorical
perception and speech production.
When interacting with another bilingual of the same 2 languages, a
bilingual will be in bilingual mode.
Here, both languages will be activated, and code switching will take place.
However, only one language will be used for processing, and is the main
language of communication
This base language will be more active than the other, which is known as
guest language
In bilingual mode bilinguals choose a base language and can bring in the
other language in a number of ways:
They can shift to that language for a word, phrase, or sentence. That is, they
can code-switch.
They can borrow a word, a short expression from other, less activated,
language and adopt it morphologically, and even phonologically, into the base
language.
Bilingual speech production
Bilingual’s language production system is dynamic.
activation states of the non-target language might differ based on certain factors
like language mode.
Colome (2001):
Phonemes :
/t/ related
/m/ cross language relation
/p/ non related
participants took longer to reject the phoneme appearing in the Spanish word than
the control one
the same pattern was replicated at different stimulus onset asynchronies (-2000,
+200, +400), which lead the authors to conclude that both the target language and
the language not in use are simultaneously activated.
This study was replicated by Hermans, Besselaar and Van Hell (2011).
The filler object names were divided into cognate and non-cognate words.
Picture: bottle
Target phonemes: /b/ affirmative condition
/f/ cross language condition
/p/ non related
The pictures were chosen in such a way that half had English names and had a
translation equivalent name in Dutch (like bottle).
The filler pictures were those whose names were also non-cognates.
The result showed that there was n difference between cross language and non-
related phoneme identification in response latency or accuracy.
Conclusion: non target language is NOT activated.
Part 2 of the experiment:
here the filler pictures were those that were cognates, like:
Moon/maan; mouse/muis etc
This time, the two critical conditions had difference in response latency as well
as accuracy.
Cross language condition took longer and were less accurate.
Conclusion: in this condition, the non-target language WAS activated.
Thus, the activation level of the non-target language can be manipulated. Its
not static but dynamic.
Similar findings have been reported by other studies too.
types
Segmental transfer:
Featural transfer
suprasegmental transfer
Phonotactic transfer
Segmental transfer:
One of the first studies investigating this was Goto (1971) and Miyawaki et al
( 1975).
Difficulty of Japanese speakers in differentiating between English liquids /r/ and /l/.
Goto used natural stimuli and tested both perception and production.
Miyawaki used synthetic stimuli and tested perception.
Both studies reported low accuracy scores for Japanese in both perception and
production.
Similar studies are carried out in recent times as well
Ingvalson, McCleland &B Holt (2011) showed that even Japanese speakers living in
English speaking environment for many years could not distinguish the two sounds.
Similar findings are reported from other studies investigating L2 specific consonants
contrasts .
Studies like this highlight the differences between L1 and L2 phonemic investory and the
resultant transfer of L1 phoneme to L2 while processing.
(Sundara, Polka &Genesee 2006; Halle, Best & Levitt 1999; Tsao, Liu & Kuhl 2006)
Featural transfer
Distinctive features like length, stress etc are used in some languages, but are
absent in others.
L2 learners ate typically found to have difficulty if these features are present in L2
but not in L1.
Length:
In some circumstances speakers whose L1 does not make use of this temporal
feature have been found to be able to perceive the same in their L2
Russian-Estonian bilinguals
English learners of Japanese find it difficult to distinguish between Japanese long and short
consonants and vowels.
Similarly, Finnish and Swedish ESL learners difficulty in perceiving the syllable final fricative /s/
Vs/z/ (peace Vs peas)
Temporal inaccuracy, location of the boundary between long and short segments etc are also
investigated.
Callan et al 2006; Han 1992; Hardison & Saigo 2010; Hirata 2004; Flege, Munro & Skelton 1992)
For example, English speaking learners of French produced longer VOTs than
French NS (native speakers).
Mandarin and Korean speakers ESL learners were found to produce longer vowel
duration than English NS and so on.
Even tonal L1 may not benefit as tones are categorized differently in different
languages.
E.g. Mandarin tone perception and production among Cantonese was studied and
results sowed the accuracy rates were not good in identifying all the tones.
(So & Best 2020; Wang, Spence, Jongman & Sereno 1999; Hao 2012)
Phonotactic transfer:
English NS producing Polish words with consonant clusters that are illegal in
English had an accuracy rate of 11% to 63%, but for those clusters that are
legal in English were 94%.
speakers of a language where consonant clusters are not allowed tend to
insert a vowel between the two consonants.
English NS living in Paris for more than 11 years were found to produce English
stops with shorter VOTs.
Chang (2012) studied American students studying in a Korean immersion program
He found general lowering of F1 and F2 values of all English vowels due to lower
similar values in Korean.