Abstract
According to their rhythmic properties, spoken languages have been classified by linguists into three categories, i.e. “stress-timed”, “syllable-timed” and “mora-timed”. Recently, this intuitive classification has been confirmed valid with perceptual studies. These studies have shown that both newborn infants and adults, upon hearing the pure rhythms of certain languages, are only able to discriminate between languages of the different categories mentioned above. These results indicate that the capacity of the categorization of the human cognitive system makes possible the classification of language that is not yet explored rhythmically. Here, we investigate the Korean spoken language through the adult’s discrimination tasks with resynthesized speech. Experimental results indicate that the Korean language is, like Japanese, one of the mora-timed languages.
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
The perception of speech rhythm is a key issue in psycholinguistics. Speech rhythm is a major component of prosody, from which intonation and phonetic-phonological cues are not included. To know the rhythmic typology of a language is important, because it seems that infants start and process language learning from the rhythm of their language. Some hypotheses argue that discrimination between languages, acquisition of syllable structure, rudimentary segmentation of fluent speech, and structuring of Head-Complement in phrases are developed through rhythm, and rhythmic typology of a language is related to such knowledge (Cutler et al. 1986; Otake et al. 1993; Jusczyk et al. 1993; Kelmer Nelson et al. 1989; Christophe et al. 2003a, b).
Infants, first, need to tune to the native language and separate it from different ones to process their native language learning, otherwise the language acquisition process does not occur. Therefore, they must distinguish the native language from other ones. Concerning the infant’s ability to discriminate between languages, several hypotheses have been proposed and tested: (a) infants might discriminate utterances of their native language from those of any other language because every language adopts its own distinctive phonetic, phonological and prosodic system (Moon et al. 1993; Bahrick and Pickens 1988); (b) infants might also distinguish between utterances of two different foreign languages for the same reasons (Mehler et al. 1988; Mehler and Christophe 1995); and (c) infants might distinguish languages according to a small number of rhythmic categories, at a stage without any a priori linguistic knowledge (Mehler et al. 1996; Nazzi et al. 1998; Ramus et al. 1999; Christophe and Morton 1998). According to several recent findings, “rhythm based language discrimination hypothesis” is surprisingly convincing among these hypotheses.
The idea of classifying languages into “stress-timed”, “syllable-timed” and “mora-timed” categories according to different impressions of timing have been proposed by linguists (Lloyd James 1940; Pike 1945; Abercrombie 1967; Ladefoged 1975). “Timing” in speech refers to the rhythmic qualities. Basically, there are three ways to assign time units in any given language: to each syllable, or to each mora, or to each stressed syllable. In syllable-timed languages, every syllable has roughly the same duration of time, while in mora-timed languages, every mora does. In stress-timed languages, every syllable or mora does not constitute the same temporal unit, but there is roughly the same duration of time between two consecutive stressed syllables. Linguists have previously hypothesized “isochronic units” as a provocation that causes distinctive impressions of these linguistic rhythms (Pike 1945; Abercrombie 1967; Catford 1977; Kiparsky 1979; Selkirk 1980; Lehiste 1972; Shockey et al. 1972; Kozhevnikov and Chritovich 1965), but measurements have failed to find the physical isochronic units (Classe 1939; Bolinger 1965; De Manrique and Signorini 1983; Wenk and Wioland 1982) and these classifications have been abandonned.
However, even if researchers have failed to find real and objective isochronic units in speech, the nature of the inexact perception of the human cognitive system in relation to real isochrony, like “perceptual center effect” (Allen 1975; Lehiste 1977; Morton et al. 1976; Fowler 1979) and the phonological phenomena to keep temporal regularity within words, like “compression” (Dasher and Bolinger 1982; Dauer 1983), are newly considered to construct such rhythms. However, decisive confirmation of the existence of such rhythmic categories has also been made recently. The recent findings of infants’ behavior concerning the discrimination between languages have been based on their categorical perception (Nazzi et al. 1998; Ramus et al. 1999). It is evident that rhythm categories exist.
However, there have been no firm agreements among linguists as to the rhythm topology of Korean. Some researchers considered it syllable-timed, others stress-timed (Han 1964; Ko 1988; Ji 1993; Park 1990). However, that it is mora-timed has not been proposed. The suggestions have not been based on perceptual research, but have relied solely on theoretical hypothesis.
We can find the rhythm category of Korean speech by observing adults’ discrimination behaviors in relation to the pure rhythms between Korean–Italian (syllable-timed), Korean–English (stress-timed), and Korean–Japanese (mora-timed) language pairs because, according to recent psychological findings, the human cognitive system, including both infants and adults, distinguishes language between different rhythmic categories only when people are tested by hearing only the pure rhythms.
Here, we present perceptual experiments not only to confirm the existence of rhythm classes, but also to find the rhythm typology of Korean speech.
Experiment
Construction of material
The method developed by Ramus et al. (1999) was used in the present research. Sentences were constructed, recorded, and digitalized at 16 kHz for the present study in four languages (English, Italian, Japanese, Korean). Four speakers (two men, two women) per language participated for the reading, where every speaker read five sentences: 80 sample utterances in total. The sentences were short declarative statements, and languages were cross-matched by the number of syllables (about 20) and the average duration (about 3 s).
The 80 sentences were segmented as precisely as possible into consonants and vowels using the sound editing software PRAAT, using both auditory and visual cues. Glides (/w/) were treated as consonants in the pre-vocalic position in a syllable, whereas treated as vowels in the post-vocalic position.
Then we measured the duration of vocalic and consonantal intervals, according to the assumption that infants distinguish only vowels (‘energy’) from consonants (‘obstacle’). For example, the phrase ‘il mio amico’ is segmented as the following: /i/ /lm/ /ioa/ /m/ /i/ /c/ /o/.
The resulting durational information was fed into the software MBROLA for synthesis by concatenation of diphones using a French diphone database. The French sounds were chosen in order to keep our discrimination tasks neutral. To remove phonetic-phonological cues, we have transformed all sentences into forms of “sasasa”, consisted of replacing all consonants with /s/, and all vowels with /a/. Then we have synthesized all these “sasasa” sentences into the form of “flat sasasa” with a constant fundamental frequency at 230 Hz, for the purpose of eliminating intonations. In this way, we have created pure syllabic rhythms of the original sentences (see Appendix).
Procedure
The AAX experimental paradigm developed by Ramus et al. (2003) was used. For each language pair (Korean vs. English, Korean vs. Italian, Korean vs. Japanese), the subjects participated in 20 oddball tasks (pilot experiments composed of the AAX paradigm, where A is one language, X is another language). We have adopted the AAX paradigm, where the first two sentences (of the same language) are formed as a context, and the third sentence is presented in either the same or a different language. After listening to each group of AAX sentences, the subjects indicated “yes” or “no” to the question “Is the third sentence (X) expressed in the same language as the previous two (AA)?”
For each language pair, the AA sentences (context) were taken from the first two speakers (speaker 1 and speaker 2 mentioned in the Appendix) of the four languages in random order and the X sentence (test) was taken from one of the last two speakers (speaker 3 and speaker 4) in random order. There were 20 trials for each language pair, and each trial (sequence of AAX) contains pauses of 500 ms between A, A and X.
Participants
Forty adult subjects (20 Koreans who did not know Italian and 20 Italians who did not know Korean) participated in each experiment.
Results
The “Signal Detection Theory” method was applied to verify the discrimination between languages. The correct percent scores were thus converted into hit rates (the proportion of “correct same” trials) and false alarm rates (the proportion of “incorrect different” trials). Hit and false alarm rates were then converted to A′ (discrimination scores), that varies between zero and one, with chance level of 0.5.
A′ scores were calculated according to the following formula:
Here, H is the hit rate and F is the false alarm rate.
Results of language discrimination experiments are as follows:
Flat sasasa | A′ | SD | P* |
---|---|---|---|
Korean–Italian | 0.70 | 0.155 | <0.001 |
Korean–English | 0.79 | 0.234 | <0.001 |
Korean–Japanese | 0.47 | 0.065 | 0.398 |
A′ scores in the first two cases confirm that rhythmic differences between Korean and Italian, a typical syllable-timed language, and between Korean and English, a typical stress-timed language, are significantly perceivable. In contrast, the A′ score in the third case confirms that results between Korean and Japanese, a typical mora-timed language, are near to the chance level, indicating that Korean belongs to the same rhythmic category as Japanese.
Standard deviation indicates homogeneity of the response values in the respective pairs of languages, in which the homogeneity in the case of Korean–Japanese is higher than the cases of Korean–Italian and Korean–English.
The probability that the sentences of the two languages presented belong to the same language is rejected in the first two cases (P<0.001) whereas, in the third case, the probability is accepted (P=0.4).
As we can see, the subjects considered the differences within the language pairs “Korean and English” and “Korean and Italian” as “above chance level”, but failed to discriminate the language pair “Korean–Japanese”. These results indicate that Korean is neither a stress-timed language like English, nor a syllable-timed language like Italian, but is a mora-timed language like Japanese.
Discussion
In this study, we have investigated Korean, in which we have found the rhythmic typology of the language to be mora-timed like Japanese, confirming the existence of such categories that have been classified traditionally. We have also found that adults can categorize languages in the same way as infants by listening to the stimuli of pure syllabic rhythms. With numerous tests, the validity of the existence of the rhythmic category of Korean was recognized, in comparison with English (stress-timed), Italian (syllable-timed), and Japanese (mora-timed). Differences in rhythm were found between English and Korean as well as between Italian and Korean. The result of not finding any difference between Japanese and Korean confirmed their unity in rhythm typology (mora-timed). We think that the present research might be a useful method for categorizing other languages not yet analyzed for their rhythm typology. Such knowledge is rather important, for previous studies have suggested that the human cognitive system can categorize languages into a limited number of categories on the basis of rhythm perception. Finally, because the even control of speech rates and accurate measurements of the phonemes were not fully made in our experiment, we think that such cases should be improved for future researches.
References
Abercrombie D (1967) Elements of general phonetics. Aldine, Chicago
Allen GD (1975) Speech rhythm: its relation to performance and articulary timing. J Phon 3:75–86
Bahrick LE, Pickens JN (1988) Classification of bimodal English and Spanish language passages by infants. Infant Behav 11:277–296
Bolinger D (1965) Pitch accent and sentence rhythm. Forms of English: accent, morpheme, order. Harvard University Press, Cambridge
Catford J (1977) Fundamental problems in phonetics. Indiana University Press, Bloomington
Christophe A, Morton J (1998) Is Dutch native English? Linguistic analysis by 2-month-olds. Dev Sci 1:215–219
Christophe A, Guasti MT, Nespor M, van Ooyen B (2003a) Prosodic structure and syntactic acquisition: the case of the head-direction parameter. Dev Sci 6:211–220
Christophe A, Nespor M, Guasti MT, van Ooyen B (2003b) Prosodic structure and syntactic acquisition: the case of the head-direction parameter. Blackwell, Oxford
Classe A (1939) The rhythm of English prose. Blackwell, Oxford
Cutler A, Mehler J, Norris D, Segui J (1986) The syllable’s differing role in the segmentation of French and English. J Mem Lang 25:385–400
Dasher R, Bolinger D (1982) On pre-accentual lengthening. J Int Phon Assoc 12:58–69
Dauer RM (1983) Stress-timing and syllable-timing reanalyzed. J Phon 11:51–62
De Manrique B, Signorini A (1983) Segmental duration and rhythm in Spanish. J Phon 11:117–128
Fowler CA (1979) Perceptual centers in speech production and perception. Percept Psychophys 25:375–388
Han MS (1964) Duration of Korean vowels. Studies in the phonology of Asian language 2 (1), University of Southern California
Ji MJ (1993) The duration of sounds. Korean Lang 3(1):39–57
Jusczyk PW, Cutler A, Redanz NJ (1993) Infants’ preference for the predominant stress patterns of English words. Child Dev 64:675–687
Kelmer Nelson DG, Hirsh-Pasek K, Jusczyk PW, Wright-Cassidy K (1989) How prosodic cues in motherese might assist language learning. J Child Lang 16:55–68
Kiparsky P (1979) Metrical structure assignment is cyclic. Linguist In 10:421–442
Ko DH (1988) A spectrograrhical investigation of vowel duration in Korean. Paper given at the memorial of Yoo Mok Sang, pp 51–62
Kozhevnikov VA, Chritovich LA (1965) Speech: articulation and perception. Joint Publications Research Service 30, US Department of Commerce, Washington
Ladefoged P (1975) A course in phonetics. Harcourt Brace Jovanovich, New York
Lehiste I (1972) The timing of utterances and linguistic boundaries. J Acoust Soc 51–56
Lehiste I (1977) Isochrony reconsidered. J Phon 5:253–263
Lloyd James A (1940) Speech signals in telephony. London
Mehler J, Christophe A (1995) Maturation and learning of language during the first year of life. In: Gazzaniga MS (ed) The cognitive neurosciences. MIT Press, Cambridge, pp 943–954
Mehler J, Jusczyk P, Lambertz G, Halsted N, Bertoncini J, Amiel-Tison C (1988) A precursor of language acquisition in young infants. Cognition 29:143–178
Mehler J, Dupoux E, Nazzi T, Dehaene-Lambertz G (1996) Coping with linguistic diversity: the infant’s viewpoint. In: Morgan JL, Demuth K (eds) Signal to syntax: bootstrapping from speech to grammar in early acquisition. Lawrence Erlbaum Associates, Mahwah, pp 101–116
Moon C, Cooper RP, Fifer WP (1993) Two day old infants prefer native language. Infant Behav 16:495–500
Morton J, Marcus S, Frankish C (1976) Perceptual centers (P-centers). Psychol Rev 83:405–448
Nazzi T, Bertoncini J, Mehler J (1998) Language discrimination by newborns: towards an understanding of the role of the rhythm. J Exp Psychol P 24(3):756–766
Otake T, Hatano G, Cutler A, Mehler J (1993) Mora or syllable? Speech segmentation in Japanese. J Mem Lang 32:258–278
Park JH (1990) The experimental phonetic study of the rhythm Korean words in view of duration. Seoul National University, MA
Pike KL (1945) The intonation of American English. University of Michigan Press, Ann Arbor
Ramus R, Nespor M, Mehler J (1999) Correlates of linguistic rhythm in the speech signal. Cognition 73:265–292
Ramus F, Dupoux E, Mehler J (2003) The psychological reality of rhythm classes: perceptual studies. In: Paper presented at the 15th international congress of phonetic sciences, Barcelona, 3–9 August 2003, pp 337–342
Selkirk E (1980) The role of prosodic categories in English word stress. Linguist In 11:563–605
Shockey L, Gregorski R, Lehiste I (1972) Word unit temporal compensation. Working Papers in Linguistics 9, Ohio State University, Colombia
Wenk BJ, Wioland F (1982) Is French really syllable-timed? J Phon 10:193–216
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
Stimuli used in experiment
Korean stimuli
Speaker 1
Kunyonun mechu hwayoilmada sengsontuikim yorirul hechunta
Menyon chirwuori omyon kwangchangeseo dongnechukchega yolinda
Shiwun kyoljongi oryowun kyoljongbota bandushi choun keosun anida
Ibon yorumun chongmal kyondiki himdul chongdoro mudopta
Kunun ibon yorumhjukatte bukyuropul yohenghal yechongida
Speaker 2
Onoyonku shilhumsili dehakneye soliptweotta
Ainun chashini kurin namuwuie sekchirul hago issotta
Kunun tekue karyoko achimilchikbuto sodulotta
Sakonun iche kwahakchokin yonkuka kanunghan desangi tweotta
Bulkwa bengnyon saie chikuye hwankyongun wanchoni talchotta
Speaker 3
Yuadulun onosupduk chokie ritumul sayonghanda
Menyon banghakimyon manun haksengduri iterirul channunda
Onoye kuchonun teo manun chungwidullo bunsokdwenda
Onul chonchasanggae kaso compyutto handaerul kuiphetta
Chukkukyongkirul boryoko manun iduri kwangchange moyotta
Speaker 4
Hankuke kyongchesanghwangi mani hochontweko issumnita
Olhenun hankuke tetongryong sonkeoka innun heimminita
Kunun achimbuto chonyok nutkekkachi chekkwa ssirumhetta
Chonyokimyon kohyange keshin bumonim sengkaki namnita
Kumnyun mare se bonyoksoka chulgandwel yechongimminida
Italian stimuli
Speaker 1
Non esiste un’unica risposta corretta a questa domanda
La maggior parte delle parole prodotte dai bambini è nuova
I turisti sono attratti dal buon clima italiano
I sostenitori della pace si sono trovati sulle piazze
Sono andato dal meccanico per i vari problemi della macchina
Speaker 2
Ho bisogno di frutta e verdura per una settimana
Il linguaggio consente di comunicare con gli altri
Questa è una finestra tipica dell’architettura gotica
Il cervello controlla più direttamente i pensieri
E’ necessaria una spiegazione alternativa a questa domanda
Speaker 3
Lui applica questi principi in un metodo d’insegnamento della limgua
Fece importanti scoperte sulla struttura del corpo umano
L’insegnante gli ha insegnato il significato delle responsabilità
La statistica permette di analizzare la scienza sperimentale
La tv ha comunicato questa mattina le notizie sull’Iraq
Speaker 4
Mario si è svegliato presto per andare a Roma
I vicini di Carla hanno deciso di cambiare casa
Essere troppo frettolosi non dà risultati positivi
Viaggiare stimola e arricchisce la mente e lo spirito
Il mercato si svolge ogni giovedi mattina in centro
English stimuli
Speaker 1
He was a very different person from my maternal grandfather
The doctor advised me to do some physical exercise
During our stay we got acquainted with a very nice couple
It was dangerous to drive under such weather conditions
London is one of the leading world centres for drama
Speaker 2
I guess I’ll ask the receptionist to bring it to my office
I’d like to reserve a table for three for 7.00 this evening
I think we might miss our train because of all this traffic
A man is looking up some information in the reference section
First report suggests engine failure could be the blame
Speaker 3
I may be going to South America for the convention next week
Acoustic cues in a speech signal include durational boundaries
He applied these principles in a method of language teaching
The German and English phonological systems are very similar
The teacher taught him the meaning of responsibility
Speaker 4
Edison was one of the most famous inventors of all time
It was a machine that can be used to record sounds
Edison helped to improve some inventions that already existed
Most Italians are very conscious of their regional origins
I thought the trip was postponed until the end of the year
Japanese stimuli
Speaker 1
Ontankabooshieno juuminno kanshimmo takamaru
Kanjano nanawaringa jibunno karuteo mitaito omotteiru
Seihuwa kisekangwasushin sankanenkekakuo kimeta
Atarashii inichio hukikomunowa horitskano sekimudearu
Seikenkootainga minshutekina chokusezsenkyode jizkensita
Speaker 2
Sengezkara pasocomo narai hajimetandesu
Hutarino kurashiwa motto iini narimashita
Karewa jisazshita maeni koiyu hanashio shimashita
Konyakuhatbyoshitasgungwa hongkommade kaerimashita
Okanenga naidakara taskeru kotowa dekimasen
Speaker 3
Buchoto bukanga shainno saiyono kotode hanashiteimasu
Henna kotoosru hitonga hueta kotodesu
Sakao angaruto shiroi tatemononga miemasu
Naniyoriwatashinga huannanowa jishinno tokidesu
Joshito bukanga hutono desainni zuite hanashiteimasu
Speaker 4
Kareno sbarashii engini watashitachiwa kokoroo ubawareta
Sono egawa mamonaku kangkokude kokai sareruyoteta
Sono ieno netaino miteo owaztamei kigadeta
Hitono hanashio yokukika nai kotodesu
Josenga hitoride kurai kohenni irukotodesu
Rights and permissions
About this article
Cite this article
Moon-Hwan, C. Rhythm typology of Korean speech. Cogn Process 5, 249–253 (2004). https://doi.org/10.1007/s10339-004-0023-1
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10339-004-0023-1