Next Article in Journal
The Effect of High-Intensity Interval Training (HIIT) on Brain-Derived Neurotrophic Factor Levels (BNDF): A Systematic Review
Next Article in Special Issue
Linguistic Prediction in Autism Spectrum Disorder
Previous Article in Journal
The Stress and Adversity Inventory for Adults (Adult STRAIN) in Korean: Initial Validation and Associations with Psychiatric Disorders
Previous Article in Special Issue
Schizophasia and Cognitive Impairment in Schizophrenia: A Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Hypothesis

Syllable as a Synchronization Mechanism That Makes Human Speech Possible

Department of Speech, Hearing and Phonetic Sciences, Division of Psychology and Language Sciences, University College London, Chandler House 2 Wakefield Street, London WC1N 1PF, UK
Brain Sci. 2025, 15(1), 33; https://doi.org/10.3390/brainsci15010033
Submission received: 10 November 2024 / Revised: 25 December 2024 / Accepted: 28 December 2024 / Published: 30 December 2024
(This article belongs to the Special Issue Language, Communication and the Brain)

Abstract

:
Speech is a highly skilled motor activity that shares a core problem with other motor skills: how to reduce the massive degrees of freedom (DOF) to the extent that the central nervous control and learning of complex motor movements become possible. It is hypothesized in this paper that a key solution to the DOF problem is to eliminate most of the temporal degrees of freedom by synchronizing concurrent movements, and that this is performed in speech through the syllable—a mechanism that synchronizes consonantal, vocalic, and laryngeal gestures. Under this hypothesis, syllable articulation is enabled by three basic mechanisms: target approximation, edge-synchronization, and tactile anchoring. This synchronization theory of the syllable also offers a coherent account of coarticulation, as it explicates how various coarticulation-related phenomena, including coarticulation resistance, locus, locus equation, diphone, etc., are byproducts of syllable formation. It also provides a theoretical basis for understanding how suprasegmental events such as tone, intonation, phonation, etc., are aligned to segmental events in speech. It may also have implications for understanding vocal learning, speech disorders, and motor control in general.

1. Introduction

Although nearly everyone can identify syllables, almost nobody can define them.
—Ladefoged (1982, p. 220) [1]
The human motor apparatus … comprises more than 200 bones, 110 joints and over 600 muscles, each one of which either spans one, two or even three joints. While the degrees of freedom are already vast on the biomechanical level of description, their number becomes dazzling when going into neural space.
—Huys (2010, p. 70) [2]
One of the fundamental problems in neuromotor control, as first recognized by Nikolai Bernstein [3], is that of too many degrees of freedom (DOF). That is, most motor movements involve multiple body structures, so that it would be immensely difficult for the central nervous system to control them separately. Speech, also as a motor skill, faces the same problem. To say a simple syllable like [ma], for example, multiple articulatory gestures need to be made concurrently: closing the glottis and increasing the lung pressure to generate voice, closing the lips and lowering the velum to produce the nasal sound [m], lowering the jaw, lowering and retracting the tongue body to produce the vowel [a] [4]. And, if the [ma] is said with a lexical tone in a language like Mandarin, the vocal folds have to be adjusted to raise or lower F0, often more than once, within the same syllable [5]. All of these need to be completed in less than one-fifth of a second [6]. In fact, it would be hard to imagine how any motor action involving more than one body structure and/or muscle would be possible without a solution to this DOF problem, and how speech would be possible without a means to reduce DOF to the extent that multiple articulators and the muscles driving them can be effectively controlled.
This DOF problem, also known as the Bernstein problem, may emerge at any level of neuromotor control, and there have been many proposals in the motor control literature on how to resolve it [7]. Bernstein’s own proposal is that motor redundancy can be minimized by freezing many of the degrees of freedom by organizing a group of muscles into a functional unit called synergy, “such that a central control signal jointly and proportionally activates all muscles in the synergy” (see p. 278 [7]). A further idea is that groups of muscles may form coordinative structures that act together to perform a single action [8,9]. Coordinative structure has also been adopted in some theories of speech production [10,11]. In particular, it is applied in the task dynamic (TD) model of speech articulation [11], which forms the computational basis of articulatory phonology (AP) [12]. In this TD/AP framework, articulatory gestures are assumed to form autonomous coordinative structures, which are then temporally overlapped with each other. To coordinate these overlapping gestures, a coupled oscillator model of timing planning is then applied [13,14,15]. In this model, each gesture is associated with an internal planning oscillator responsible for controlling the temporal pattern of its movement. In this conceptualization, however, not only the multiple gestures, but also their relative timing, are separately controlled. This would introduce more rather than less degrees of freedom.
The coupled oscillator model has been used to account for various synchronization phenomena in motor movements [16,17,18] because motor synchrony appears to resemble the well-known physical entrainment [19], as both show the shared timing of two oscillating movements. Entrainment is a physical phenomenon whereby two oscillating systems with similar natural frequencies, e.g., two pendulum clocks, gradually fall into synchrony when they are connected through some mechanical link, such as being hung on the same beam [19,20]. If coupled oscillation does not resolve the DOF problem as pointed out above, however, we may wonder whether entrainment is the right analogy for motor synchrony. Indeed, a careful comparison reveals that motor synchrony differs from entrainment in a number of critical ways, as listed in Table 1. First, motor synchrony can occur in bi-manual actions with no repeating cycles [21]. Such non-repeating synchrony, by definition, would be irrelevant to entrainment. But it is highly relevant for monosyllabic words spoken in isolation. Likewise, in speaking in unison—a skill surprisingly natural to most people without much practice, speakers can easily synchronize their reading aloud of the same text [22]. The non-periodic articulatory movements in speaking in unison cannot be accounted for by theories that use periodicity as the basis for explaining synchronization [23].
Second, in entrainment, it takes many cycles for two oscillators to reach synchrony. In motor synchrony, the shift from 180° (anti-phase) to 0° (in-phase) occurs in only 1–2 cycles [24,25,26,27], which is virtually instantaneous. In a system of coupled oscillators, the fastest phase shift [16] simulated with an oscillation model takes 5–6 cycles to complete. A gradual shift across 5–6 cycles also means that in some of those cycles, the phase relation is neither 180° nor 0°, which is exactly what has been repeatedly shown to be impossible in motor synchrony [26,27,28]. Furthermore, in coupled oscillation, each oscillator has to have its own initial phase condition. For speech, one would naturally ask, where do the initial phase conditions come from in the first place?
Third, as shown in the third and fourth rows of Table 1, entrainment requires that the synchronized oscillators are similar in their natural frequencies, and even after reaching synchrony, they may go out of phase again [29,30]. Neither high similarity in frequency nor phase instability is characteristic of motor synchrony [21,26].
Finally, probably the most fundamental difference is that in entrainment, the systems being synchronized are independent of each other, with no central control, and the synchrony is achieved passively through physical links between the involved systems [20]. Motor synchrony, in contrast, occurs between movements that are under a common central control, or in the case of synchrony between two individuals, under a shared control maintained through sensory monitoring [27]. The central control, as well as the sensory monitoring that makes it possible, is clearly lacking in physical entrainment.
The Bernstein problem therefore is unlikely to be solved by a coordinative structure constructed as a system of coupled oscillators [13,14,15]. To start with, there is no explicit scheme to reduce degrees of freedom. Rather, each of the assumed planning oscillators has to have its own natural frequency unrelated to the natural frequency of the associated gesture, and its own initial phase condition, which results in at least two additional degrees of freedom. Furthermore, entrainment is used to model both abrupt shifts from VC to CV in accelerating repetitive syllable sequences [25], and the planning of intergestural coordination is postulated to occur before the onset of each and every syllable [13,14,15]. It is especially problematic that it takes at least several cycles to complete a phase shift or to stabilize the phase relation during planning when they are computationally modeled as an entrainment process [16].
A viable alternative solution therefore needs to explicate how degrees of freedom can be effectively reduced. The solution considered in this paper is that for speech, the syllable serves exactly this function. This solution, however, suggests a theory of the syllable that differs from all past syllable theories. And it also touches on the long-standing issue of coarticulation. The following Section briefly reviews the state of the art for both syllable and coarticulation.

2. Syllable and Coarticulation

That speech utterances are made up of syllables may seem obvious. Most, if not all, early writing systems (Sumerian, Linear B, Akkadian cuneiform, Chinese, Mayan, etc.) started as syllabaries, in which the written symbols represent syllables (or sometimes morae) rather than consonants and vowels [31,32,33]. It is also much easier for anyone, including non-experts, to count the number of syllables in a word than the number of segments in a syllable [33,34,35]. The syllable is also known to play many important roles in speech. It is the unit that carries stress and accent [36,37,38], rhythm [39,40,41], and tone [42,43]. It is said to be a unit in a prosodic hierarchy of constituents [44,45], which influences patterns of segmental allophony, and a unit that serves as the domain of applying language-specific phonotactic patterns and other phonological rules [46,47]. It is also critical for the perceptual segmentation of the speech signal [48,49,50]. However, neither our intuition about it nor its own usefulness has been sufficient to avert doubts about the syllable. After examining eight lines of traditional evidence in support of the syllable as a representation unit in speech production, Shattuck-Hufnagel in 2011 found none of them unequivocal [35]. The lack of clear evidence has led to skepticism about its existence [51] or universality [52]. Similar reservations have been expressed by Gimson [53], Steriade [54], and Blevins [55]. A major reason behind these doubts is that the nature of the syllable has remained vague. In particular, no theory has been able to address some of the hardest questions:
  • Why are there syllables?
  • Do syllables have clear phonetic boundaries?
  • Do segments have definitive syllable affiliations?

2.1. Why Are There Syllables?

For any scientific inquiry, the why question is often the most essential yet the most difficult. This is also the case with the syllable. The functional use of the syllable, as mentioned above, to carry stress, accent, tone, rhythm, or to serve as a domain of phonological rules, are all extended benefits of the syllable, and so cannot be the primary reasons for the existence of the syllable in the first place. What we are seeking is an account of why the syllable is indispensable, i.e., serving a function that is so vital that speech would be impossible without it.
Some theories have taken the syllable as the basic unit of speech, e.g., Stetson’s motor phonetics [56] and Fujimura’s C/D model [57]. But they have offered no explicit proposal as to why syllables are obligatory at the articulatory level. In MacNeilage’s frame/content theory [58], the syllable is suggested to have evolved from the oscillation of the jaw in movements like chewing, sucking, and licking. However, the ability to oscillate the jaw is shared by virtually all mammals, yet not even our closest relatives, i.e., chimpanzees, gorillas, and bonobos, have developed syllable-based speech [59,60]. Thus, being able to oscillate the jaw does not seem to inevitably lead to an ability to articulate syllables. Something extra must be involved.
It has also been proposed that the syllable is a unit of stored motor programs [61,62]. But the proposal is questioned for its inability to explain cases of resyllabification or the lack thereof [35]. More importantly, even if stored syllable-sized motor programming is shown to exist, it cannot explain why the unit has to have the form of a syllable. It thus remains an open question as to whether the syllable, with its own unique characteristics, is indispensable, i.e., serving a function that is so vital that speech would be impossible without it.

2.2. Are There Clear Boundaries to the Syllable?

Given an utterance like the one shown in Figure 1, it may seem that some of the syllables are well separated by the alternation of consonants and vowels whose spectral patterns show clear boundaries [63]. However, the syllable boundaries are much less clear-cut in the case of /wei/. Because it begins with a glide /w/, it is hard to determine when the preceding syllable ends and the next one starts. Even more difficult are cases where a word starts with a vowel, as in the English words like artist, article, articulate, and arbitrary. When they are preceded by words ending in a vowel, as in new artist, my article, to articulate, or fairly arbitrary, there would be continuous formant movements across the word (hence, syllable) boundaries (unless when spoken very carefully so that the syllable starts with a glottal stop). The same problem would be seen in cases of word-internal syllables, like in hiatus, appreciate, mediocre, etc., where there should presumably be a syllable boundary between /i/ and the following vowel or diphthong, yet all we can see in the spectrogram in most cases are continuous formants between the preceding and following consonants.
The difficulty of syllable boundary identification has led to the view that it is simply futile to look for clear-cut boundaries in the speech signal, as argued by Hockett [66], who likens segments to colored raw Easter eggs lined up on a belt. After being crushed by a wringer, the heavy smearing makes the edges of the individual eggs unrecognizable. The problem with this analogy is, however, if we do not know where the boundaries are, how can we be so certain that segments are heavily overlapped with each other [67,68]? So, the fuzziness of the syllable boundaries is directly related to the fuzziness of segmental boundaries, which in turn is related to yet another major conundrum of human speech: coarticulation, as will be discussed in Section 2.4.

2.3. Do Segments Have Definitive Syllable Affiliations?

The clarity of syllable boundaries hinges on not only the clarity of segmental boundaries, but also the certainty about where each and every segment should belong in a syllable: onset, offset, or between two adjacent syllables, i.e., being ambisyllabic. There have been many theories of syllabification, including the law of initials and the law of finals [69], the maximal onset theory [70,71], the theory that stressed syllables are maximized [72,73], and the weight-stress principle [73,74,75]. But so far, there has been no consensus on even some of the simplest cases. For the word happy, for example, at least four ways of syllabification are possible as summarized by Duanmu [76]: /hæ.pi/, /hæp.i/, /hæpi/, and /hæp.pi/ (where the period stands for syllable boundary and an underscore indicates the segment is ambisyllabic). All these syllabification theories, however, are based on intuition or non-experimental phonological analyses. There are also experimental investigations of syllabification intuition by naïve speakers [49,77,78,79]. None of the syllabification findings, however, has directly addressed the issue of what syllable boundaries look like in the acoustic signal or in terms of articulatory movements.

2.4. Are Syllables Related to Coarticulation?

The problems of syllable boundary and syllable affiliation of segments discussed above are both closely related to another long-standing problem, namely, coarticulation. The term coarticulation, initially “Koartikulation” in German, was coined to refer to articulatory timing around syllable onset [80]. The observation was that “the articulatory movements for the vowel in tokens such as /ma/ or /pu/ began at the same time as the movements for the initial consonant” ([68] p. 14). The link between syllable and coarticulation is further strengthened by Kozhevnikov and Chistovich [81], who proposed the notion of articulatory syllable, based on the observation that in Russian, the lip protrusion of /u/ begins at the same time as the first consonant in a consonant cluster. According to this notion, the domain of coarticulation is the articulatory syllable, in the sense that all the articulatory actions connected with the articulatory syllable, including the vowel, start at the syllable onset, as long as the consonantal movements do not contradict the articulation of the vowel.
The articulatory syllable, however, has been questioned due to uncertainties over the temporal scope of vowels in a syllable [82,83,84,85]. A major reason for the skepticism is the widely reported preparatory activities, particularly the classic finding of Öhman [86], and the phenomenon of vowel harmony [87], which seem to suggest that the scope of the vowel extends well before the syllable onset. By now, the term coarticulation is generally used to refer to virtually any variability of a segment due to the influence of surrounding segments [68].
Perhaps the most detailed account so far of coarticulation as related to the syllable is offered by the TD/AP framework [13,14], as mentioned earlier. In the most recent version of TD/AP, syllable structure is modeled as emerging from coupled oscillations as internal planners that are in-phase between consonant and vowel gestures at the syllable onset but anti-phase at the syllable offset. The in-phase coupling of CV at the syllable onset is consistent with the notion of articulatory syllable [20], and accounts for a large amount of CV coarticulation. However, the TD/AP account of the syllable leaves some core problems unresolved. The most critical is the assumption that each gesture is controlled by a planning oscillator whose frequency and initial phase both need to be specified, making it unclear how DOF can be effectively reduced. Second, the in-phase and anti-phase assumption for syllable onset vs. offset is based on empirical observations, but there is no account of why the asymmetry is there in the first place. Finally, as shown in Table 1, it is questionable that coupled oscillation based on physical entrainment is the right model for motor synchrony due to multiple discrepancies between the two phenomena.

3. Syllable as a Synchronization Mechanism

The hypothesis considered in this paper is that the DOF problem is solved by actively controlled motor synchrony, which, in the case of speech, is achieved through the syllable. This synchronization fixes the relative timing of multiple motor movements so that most of the temporal degrees of freedom are eliminated, not only in learning, but also in normal operation (as opposed to Bernstein’s proposal that the freezing of DOF is mainly for learning). For speech, the formation of the syllable is also the mechanism underlying coarticulation. The syllable model based on this hypothesis consists of three core mechanisms: target approximation, edge synchronization, and tactile anchoring, as sketched in Figure 2. Target approximation (the dashed curves) is the articulatory process of executing phonetic targets, and it is what defines the articulatory gesture (see Section 3.1 for details). Edge synchronization (indicated by the vertical lines) is the mechanism of coordinating multiple gestures that make up a syllable, including consonant, vowel, tone, and phonation. (Here, phonation refers to the use of voice quality as an independent dimension to mark lexical contrasts, which is found in some languages [88,89]. It does not refer to phonation properties that accompany consonant manner of articulation). And tactile anchoring (not directly represented in Figure 2) is the facilitation of edge synchronization by sensory feedback, mainly through tactile sensation during consonant closures.
Conceptually, the three mechanisms of the synchronization hypothesis are interlocked as illustrated in Figure 3: target approximation is what defines the onsets and offsets of individual movements; movement onsets and offsets (not acoustic landmarks) are what edge synchronization aligns; and tactile anchoring is what ensures the accuracy of synchronization.
By positing the syllable as a mechanism for solving the DOF problem, the synchronization hypothesis not only offers an account of the syllables that deviate from existing theories but also provides an account of coarticulation, as will be detailed in Section 3.2.1, Section 3.2.2 and Section 3.2.3. In the following sections, each of the three core mechanisms of the synchronization hypothesis will be elaborated, with support from existing literature. Also discussed will be the similarities and differences between this hypothesis and other models in addressing various specific aspects of the syllable. Finally, Section 4 and Section 5 will present new evidence obtained in the latest studies.

3.1. Targets and Target Approximation

The notion of target approximation goes back at least as far as Lindblom (1963) [90], who suggests that underlying phonetic targets are often only partially realized due to time constraints. Similar ideas are shared by a number of models proposed since Lindblom [90], in particular, the Fujisaki model of intonation [91], and the TD/AP framework for segmental articulation [11,12]. The version of target approximation presented here [92], as schematized in Figure 4, was independently developed based on empirical data on contextual tonal variations [5,93,94,95].
In this model, each movement is a process of approaching an underlying target (dashed lines in Figure 4) within an extrinsically designated temporal interval. Each target approximation movement is controlled by three parameters: target position, target slope (underlying velocity), and target strength. Adjacent target approximations are contiguous without overlap, shifting abruptly from one to the next at each interval boundary. The resulting surface contour (solid curve in Figure 4) is nevertheless smooth and continuous due to the transfer of dynamic states at the boundary. Despite similarities with other models, there are five key properties that are unique to the target approximation model presented in Figure 4:
  • Surface acoustic forms result from asymptotic approximation of a single sequence of underlying targets rather than from superposition of multiple underlying contours [91,96,97].
  • Targets are approximated sequentially, with neither overlap of adjacent movements along the same articulatory dimension [98], nor gaps in between [38,91], unless there is a silent pause. The lack of gaps also means that there are no temporal intervals (except pauses) without targets.
  • Targets can be intrinsically dynamic, i.e., with underlying slopes of various degrees. No other model, to our knowledge, has incorporated dynamic targets. (See Section 3.1.5 for critical differences between underlying velocity and surface velocity. The former is a property of the target, which can be either static or dynamic, while the latter is the consequence of executing the target. Some models, like TD and Fujisaki models, specify the stiffness of the target gesture, which indirectly specifies surface velocity. But they have no specifications for underlying velocity. So, a fully achieved target in those models can only generate an asymptote to a static articulatory state.)
  • Every target also has a strength specification, which determines the rate at which the articulatory goal is approached. Target strength (or stiffness) is treated in other models as either mostly fixed [91,98] or a means of controlling duration [99].
  • Target duration is not predominantly determined by the time needed to reach the target, or intrinsic timing [12,100], but by functional factors such as lexical contrast (lexical quantity, lexical tone, and lexical stress), focus, and boundary marking [101].
The target approximation model has been quantified in the form of qTA [102], which has been applied to English, Mandarin, Thai, Japanese, Arabic, Persian, Savosavo, Fijian, and Vietnamese [103,104,105,106,107,108,109,110,111,112]. The present paper will not focus on the quantitative aspect of the model, but some of the graphics are generated with qTA. Also, although the target approximation model was initially developed for tone and intonation, its relevance for segments has also been demonstrated [113,114,115]. Birkholz et al. [116] have developed a higher-order version of the target approximation model for an articulatory synthesizer, which has been successfully implemented in articulatory synthesis for English, Thai, and German [117,118,119,120,121].
In the following Section, the main properties of target approximation will be elaborated and evidence from the literature, wherever available, will be presented.

3.1.1. Asymptotic Approximation

Among the clearest evidence of asymptotic approximation are the F0 trajectories of lexical tones found in connected speech [5,94]. As shown in Figure 5, the F0 of the tone in the third syllable in each plot starts at very different heights depending on the tone of the second syllable. Yet all the trajectories quickly accelerate away from the initial states, and converge, within the third syllable, to a linear configuration that reflects the tone’s underlying targets: high-level, falling, and rising, respectively. Similar asymptotic approximation has also been observed for vowels [122].

3.1.2. Sequentiality

In Figure 4, although the surface trajectory is smooth and continuous, the underlying targets are strictly sequential, with neither gap nor overlap around the boundary. Thus, there is no need for specifications (hence, no extra degrees of freedom) on how much adjacent targets along the same articulatory dimension overlap with each other, or whether a temporal interval is targetless. But there are also alternative conceptualizations on the sequencing of targets. One is gestural blending and the other is intermittent target specifications. Gestural blending is seen in articulatory phonology [12], which assumes that gestures can temporally blend with each other even for the same articulator. Gestural blending is used to explain anticipatory coarticulation as well as undershoot [12,123]. The execution of gestural blending is implemented in the task dynamic model as weighted averages of the blended gestures [12]. There is evidence, however, that the movements of any single articulatory dimension result from sequential rather overlapping execution of successive targets. This is shown for tongue body [124], velum and lips [125,126], and f0 [127]. Also, Ostry et al. [128] have demonstrated that a model based on the equilibrium point (EP) hypothesis of motor control [129] is able to generate kinematic movements that show coarticulatory overlap with non-overlapping underlying control signals. Here, a conceptual difficulty is the non-unique relations between the observed articulatory/acoustic trajectories and possible underlying control parameters, as illustrated in Figure 6.
In panel (A) of Figure 6, there are three successive target approximation movements, each largely attaining its target by the offset. These movements are strictly sequential, as indicated by the alteration of the line patterns and colors. Panel (B) also shows three target approximation movements, but the first one is shortened relative to the first movement in panel (A), resulting in undershoot, i.e., an incomplete attainment of the target. From the graph, it is clear that the undershoot is due to a premature termination of the first movement by the early onset of the second one, which truncates the former. But the truncation also makes the offset of the first movement appear “assimilated” to the second target, as indicated by the arrow. When the time reference (vertical line) remains unchanged from panel (A), the first movement also appears to “anticipate” the second one, although there is no true anticipation given the clearly marked movement boundary. In panel (C), instead of truncation, the final portion of movement 1 and the initial portion of movement 2 are overlapped. The overlap is implemented by inserting a new target that is the average of the first and second targets. (There are also other, more sophisticated ways of blending, e.g., averaging, suppressing, and adding [11].) This blending thus explicitly models an “anticipatory assimilation”. The resulting trajectory, however, is not very different from the one in panel (B) if the boundaries are ignored. (Compared to panels A and B, the second target in panel C is less fully attained. This is because the blending also shortens the target approximation movement of the second target. Thus, there is less than enough time to reach the target even though the onset of the movement is actually higher than in the other two panels. This means that despite the similarity, different assumptions about sequential arrangements do lead to slight variations in surface trajectory, making direct computational comparisons possible in future research.) Thus, truncation can generate trajectories very similar to those generated by blending, but it has the advantage of not needing to specify the amount of overlap, thus eliminating a critical degree of freedom.
Sequentiality through truncation has a number of further implications. The first is that the duration of target approximation is mostly an extrinsic rather than intrinsic property of the gesture [100], which allows it to be specified by external functions like word stress, lexical quantity, focus, boundary marking, etc. [101]. Secondly, given the frequent occurrence of truncation due to the extrinsic factors [115] and the fact that any degree of truncation is possible (even up to 100%, e.g., in syllable contraction [113]), target approximation is rarely a 0–360° full cycle. Thus, it is inappropriate to model inter-gestural alignment in terms of phase relations such as being in-phase or anti-phase [14]. This would be a further reason, in addition to those listed in Table 1, against modeling motor synchrony as physical entrainment. Finally, the massive range of possible truncations [113,115] makes it impossible to control duration through articulatory strength [130,131], because, for example, it is inconceivable that an extreme shortening of a segment or syllable up to full elimination is achieved by a maximum increase in stiffness.

3.1.3. Full vs. Underspecified Targets

Underspecification has been a popular idea in both phonology and phonetics to account for severe undershoot or lack of apparent targets [132,133,134,135]. The hypothesis is that some units do not have fully specified phonetic values, and their surface patterns come from interpolation between adjacent, fully specified units. Boyce et al. [126] have shown, however, that intervals with highly variant lip rounding and nasalization properties may still stem from specific underlying goals, as observed with minimal contrast comparisons of articulatory movements. Similar findings have been made for the neutral tone in Mandarin [112,127], which has often been considered targetless. In Figure 7A, the F0 contours of the Falling (F) tone in the second syllable converge quickly to a falling slope following the four full tones in the first syllable. In contrast, the neutral-tone sequences in Figure 7B do not show full convergence of F0 by the end of the second syllable. But by the end of the third neutral tone in Figure 7B, all the trajectories have approached a mid-level F0. This approximation indicates that the neutral tone has its own target, which is halfway between the Falling tone and the Low tone, as evident from Figure 7C. But the slow approximation in Figure 7B, as opposed to the quick convergence in Figure 7A, suggests a weak articulatory strength in the realization of the neutral-tone target. Note that the assumption of no underspecification further reduces DOF by eliminating the need to specify for every temporal interval whether some of the target properties are missing [12,136].

3.1.4. Target Approximation vs. Its Preparation

From Figure 7C, it is also apparent that there is no anticipatory effect of the F0 differences due to the tone of the final syllable upon the preceding neutral tones, as shown in Figure 6. This suggests that there is no need to assume a leftward overlap of the full-tone target with the preceding target even if it is weak. But anticipatory preparation has been a popular idea for segmental articulation [68,137]. Yet the definition of preparation has been unclear. As an illustration, Figure 8 shows the decomposition of a badminton smash, a complex skilled motor action. The goal of the action is to strike the shuttlecock as hard as possible, which is achieved by a unidirectional arcing movement of the racket (frames 4–6). But before that, the racket is moving in the opposite direction in preparation for the main arcing movement (frames 1–4). The function of this preparatory movement is to maximize the travel distance for the racket during the smash, with the goal of achieving a high velocity. Similar preparatory movements have been shown for both singing and speech. For singing, a preparatory movement in the opposite direction from the target note is found to be a core property of the singing voice [138]. For speech, pre-low raising, which increases the pitch of a non-low tone before a low-pitched tone, has been reported for a number of languages [5,94,139,140,141,142].
In contrast to the preparatory movements in frames 1–4, the movements in frames 4–6 are all in the direction of making the ultimate contact, and so no part of it, including the initial portion, say frames 4–5, should be taken as preparatory activities separate from the rest of the arcing movement in the same direction. By the same token, the entire unidirectional movement toward a phonetic target should be considered a single-target approximation action. This distinction between preparation and target approximation will be highly relevant in the upcoming discussion of coarticulation. (It is possible, however, that even the preparatory movements in frames 3–4 of Figure 8 are part of the smash action. Whether this is the case could be determined by the timing of the movement sections. In Figure 5, e.g., the anticipatory raising movement, hence, the “preparation”, before the L and F tones seem to start from the middle of the preceding syllable. That is where the second syllable actually starts, as argued in the subsequent discussion. This possibility has been investigated in recent research, as will be discussed in Section 3.2.4.)

3.1.5. Dynamic Targets and Velocity Propagation/Continuity

From Figure 5B,C, we can see that the F0 contours of the Falling tone after four different tones all converge to a linear falling slope, and those of the Rising tone all converge to a linear rising slope. It has also been shown that in both dynamic tones of Mandarin [95] and Cantonese [144], and diphthongs in American English [145], the final velocity of F0 and formants remains largely constant when the speech rate varies from normal to slow. Thus, a specific velocity is aimed at as part of the phonetic target associated with those linguistic units. Dynamic targets are actually commonplace in other motor movements. Again, from Figure 8, we can see that when the target is reached at frame 6, what is achieved is not only a particular position of the racket, but also a high-speed impact on the shuttlecock. Thus, the target of the smash is dynamic, consisting of both position and velocity specifications. Also, given a high velocity as part of the goal of a dynamic target, its achievement may have a powerful carryover effect on the following movement. In Figure 7C, for example, the final velocity of the Rising tone is so high that the F0 rise continues for more than half of the syllable in the following neutral tone.

3.1.6. Summary of Target Approximation

There is much evidence that continuous surface movements of both articulatory and acoustic dimensions result from strictly sequential approximation of successive targets, and each approximation is executed with a specific articulatory strength. On the other hand, there are also alternative models that assume temporal overlap of auto-articulator gestures and underspecified targets, and targetless intervals. Although those models can theoretically also generate contextually variant surface trajectories, strictly sequential and fully specified targets have the advantage of assuming fewer degrees of freedom and offering a simpler basis for defining the onset and offset of articulatory gestures, which is critical for edge synchronization.

3.2. Edge Synchronization

As shown in Figure 2, edge synchronization means that (a) the onset of the syllable is the start of the target approximation for most of the syllabic components, including the initial consonant, the first vowel, the lexical tone, and the phonation register (here, phonation refers to the use of voice quality as an independent dimension to mark lexical contrasts in some languages [88,89]; it does not refer to phonation properties that accompany consonant manner of articulation.); and (b) the offset of the syllable is the end of all the remaining movements. The mechanism therefore entails full synchrony at both edges of the syllable. The synchrony is asymmetrical across the syllable, however. At the left edge, there is a synchronous onset of all the syllabic components involved, while at the right edge, there is the synchronous offset of only laryngeal components with either C or V, but not both. The benefit of synchronization is already discussed in Section 2, and the following discussion is only on the evidence and manner of its realization.

3.2.1. C-V Synchronization and Coarticulation

As mentioned in Section 2.4, a major objection to the notion of articulatory syllable [81] is that the scope of the vowel should extend well before the syllable onset based on the classic finding of Öhman [86]. That is, in a V1CV2 sequence, the activity of V2 can be seen during V1, as shown in Figure 9, where F2 starts to rise well before the closure of /b/. Öhman’s interpretation of this “anticipatory” activity is that “a motion toward the final vowel [V2] starts not much later than, or perhaps even simultaneously with, the onset of the stop-consonant gesture”. [86] (p. 165). But acoustically, the start of that activity, namely, the rise of F2, is well inside V1, which gives the impression that V1 is coarticulated with V2 [68]. Here lies, therefore, the key challenge of coarticulation: the discrepancy between the articulatory and acoustic onset of a phonetic unit, namely, articulation starts well ahead of acoustics. Now, the explicit definition of sequential target approximation in Section 3.1 would suggest that, in fact, there is no discrepancy between articulation and acoustics. This is because any acoustic movement away from the target of a sound is by definition no longer part of that sound. By the time F2 starts to turn upward in Figure 9, the articulation of V1 (/a/) is already over, and the articulation of V2 (/y/) has already begun, as illustrated in Figure 6B. There is therefore no evidence of anticipatory coarticulation of V2 with V1 from Öhman [86].
But true coarticulation in the sense of co-production [146] does happen, between syllable-initial consonant and the first vowel of the syllable. In Figure 9, roughly at the time when F2 makes an upward turn, F1 starts to go down toward the low valley in the /b/ closure, indicating that the articulation of /b/ also starts from there. What is not clear is whether the target approximations of V1 and C happen exactly at the same time, as Öhman [86] did not directly compare the timing of the articulation of C and V. And, despite the postulation of the synchronous C-V co-onset resulting from planning gestural oscillation [13], subsequent studies under articulatory phonology framework have repeatedly reported asynchronous C-V onsets [147,148,149,150,151]. This has led to the declaration that the newly accepted generalization in AP is that the vowel gesture starts somewhere after the onset of the consonant closure gesture but before the release gesture [152]. But as pointed out in Liu et al. [153], the CV asynchrony found in these studies is due to a flawed method of determining gestural onsets based on velocity profiles of individual gestures themselves, which is susceptible to confounding from adjacent and concurrent gestures and variable intrinsic stiffness of different gestures.
Clearer evidence is found in a series of studies based on a minimal triplet (or double minimal pair) method proposed in [154], which uses the gestural divergent point of C and V minimal pairs as indicators of their respective onsets, and determines CV synchrony by comparing the timing of these onsets. In Xu and Gao [155], for example, the stimuli were minimal triplets of syllable sequences in Mandarin in the form of C1V1#C2V2, as shown in Figure 10. Each triplet consists of two minimal pairs. The first contrasts the two consonants in C2: [j] vs. [l], while the second contrasts the two vowels in V2: [i] vs [u]. In the first minimal pair, the divergent point of the F2 trajectories indicates the onset of C2, because that is where the two consonants start the approximation of their respective places of articulation. In the second minimal pair, the divergent point of F2 indicates the onset of V2, because it is where the two vowels start the approximation of their respective vocal tract shapes. The two consonants are both sonorants with incomplete closure of the oral cavity, so that formant movements during the consonantal constrictions can be traced. In addition, all the words have a Rising tone on both syllables, which allows the two resulting F0 peaks to serve as time references for the onset and offset of the second syllable [156].
Figure 11 shows the mean F2 trajectories of four of the triplets in [155] produced by three male speakers. In each plot, the solid and dashed lines differ in the initial consonants: [l] vs. [j], and the point at which the two trajectories start to diverge indicates the onset of both consonants. The solid and dotted lines, on the other hand, differ in the vowels of the second syllable: [i] vs. [u], and the point at which the two trajectories start to diverge indicates the onset of both vowels. Strikingly, in each case, the vowel divergent point occurs at about the same time as the consonant divergent point. This common divergent point, as indicated by the vertical arrows in Figure 10, is well ahead of the onset of the [l] closure. This finding is recently supported by a more systematic study on Mandarin, with parallel articulatory (EMMA) and acoustic data and Bayesian statistics [153]. In addition, also using the minimal triplet paradigm, Liu and Xu [157] showed that in CCV syllables in British English, the vowel gesture starts in synchrony with the very first consonant, just like what is described for Russian [81]. (Note that this finding contradicts the C-center model, which aligns the vowel onset to the center [158]. But it is important to note that the C-center effect refers to the consistent duration of the conventional vowels in syllables with consonant clusters of various lengths. As such it does not directly assess the actual alignment of vowels with the consonant clusters.) There is therefore rather clear evidence by now that consonant and vowel target approximations start at the same time at the onset of the syllable.

3.2.2. Coarticulation Resistance

The coproduction of C and V at the syllable onset means that they would interfere with each other’s articulation because they often have conflicting goals. This would result in variations in their acoustic output. Given coproduction as the mechanism of coarticulation, however, there has to be a solution to the articulatory conflict between the coproduced phonetic components. It is known that some segments show better ability to resist coarticulatory variation [159,160,161]. A major source of such coarticulation resistance is the amount of constraint that a consonant or vowel places on the tongue body [160,161]. Those with intrinsically stronger tongue body constraints show greater resistance to coarticulatory influence than those with weaker constraints.
What can be first recognized is that the severity of the conflict would depend on the number of articulators shared by the co-produced sounds. The least conflict occurs between well-separated articulators, e.g., the larynx and the oral articulators, as will be discussed in Section 3.2.4. The most severe conflict would occur when virtually all the articulators receive clashing demands. This would happen between glides like /i/ or /w/ and the following vowel. As semivowels, their articulatory targets specify the shape of the entire vocal tract, just like a vowel. The glide and vowel targets, therefore, have to be sequentially approximated, as can be seen in Figure 10B for /w/ between /i/ and /ei/. If some of the articulators are shared while others are free to serve either of the two sounds involved, an obvious solution is for the shared articulators to sequentially approach different targets, while allowing the rest of the articulators to simultaneously approach their respective targets. In /ba/, for example, the shared articulator, the lips, can first make the bilabial closure and then open up for the vowel. At the same time, all the lingual articulators, with no positional requirement from /b/, would be free to start approaching their /a/-specific targets from the syllable onset.
The trickiest situation is when two segments share the same primary articulator, as in the case of velar consonants like /ga/ and /ka/. Because the tongue body needs to contact the velum, there is a direct conflict with the coproduced vowels, which also have clear specifications for the tongue body. How is it possible, then, for sequential target approximation to resolve the articulatory conflict? Our hypothesis, as already hinted in Section 3.1.2, is dimension-specific sequential target approximation (DSSTA); that is, sequential target approximation can be specific to individual dimensions of an articulator rather than always over the whole articulator. For the velar consonants, as the tongue dorsum needs to be raised to make the velar contact, the vertical position of the tongue dorsum therefore has to first approach the consonant target before turning to the vowel target. The precise horizontal position of the tongue body, in contrast, is probably less critical for the consonant. Therefore, the horizontal position of the tongue dorsum can start to move toward the vowel target right from the syllable onset. As a result, the point of contact between the tongue body and the palate for /k/ would naturally vary gradiently with the coproduced vowel: more advanced for front vowels, and more retracted for back vowels [162]. Conceptually, therefore, dimension-specific sequential target approximation not only resolves the problem of coarticulation resistance but also explains how CV co-onset is articulatorily implemented in general.
Two recent modeling studies have put DSSTA to the test. In [163], an articulatory synthesizer [164] in which the dynamics of all articulators are controlled by a target approximation model, was trained with acoustic signals of CVC syllables to learn articulatory targets of consonants and vowels. During learning, the training algorithm allowed tongue dorsum height to be controlled by the velar stop up until the moment of tightest closure, and the tongue dorsum frontness was controlled by the vowel from syllable onset to vowel offset. Not only was the variable velum contact location successfully learned, but also the gV syllables synthesized with articulatory parameters learned this way were highly intelligible, with mean recognition rates of 78%, 100%, and 83% for get, god, and good, respectively. In [121], the application of DSSTA was found to be effective in the control of not only tongue dorsum in velar stops, but also other articulators in bilabial, alveolar, as well as velar stops.

3.2.3. Locus and Locus Equations

Coarticulation resistance is also closely related to two other classical phenomena, namely, locus and locus equations. Locus is a phenomenon observed soon after the classical discovery that F2 transitions carry perceptual cues for consonant place of articulation [165]. It was found that many of the transitions point back in time to a locus such that as long as the first part of the transition is silent, the same consonant is heard [166,167], as shown in B in Figure 12a for /d/. The presence of the entire transition would, in contrast, result in hearing different consonants depending on the following vowel (A in Figure 12a). As reasoned by Delattre et al. [166] (p. 772), the locus phenomenon indicates that “no appreciable sound is produced until at least part of the articulatory movement has been completed”. But it is left unexplained why the early part of the transition movement has to be silent.
Closely related to the locus phenomenon are locus equations [88,169]. Locus equations refer to the finding that the onset of vowel F2 transition after a given stop consonant (i.e., the equivalent of the second vertical dash in Figure 12b(B)), when plotted over F2 at the “center” of the vowel (equivalent of the plateau in Figure 12b(B)) across different vowels, shows a strong linear relation. Fowler [170] and Iskarous et al. [171] have linked the linearity in the locus equations to coarticulation resistance, arguing that it demonstrates the invariance in coarticulation resistance across different vowels. Lindblom and Sussman [169] have further linked the locus equation back to the classical locus phenomenon, proposing that the critical articulation of a stop consonant is the target: lips for /b/, tongue blade for /d/, and tongue body for /g/, but the rest of the articulators have no specified target and so are allowed to be coarticulated with the vowel. This account comes very close to the coarticulation resistance account by Fowler and colleagues. Common to both accounts, however, is that they are concerned only with the articulatory or formant movement from the voice onset to the center of the vowel. Lindblom and Sussman ([169] p. 17) explicitly assumed that “the movements of the articulators in a CV syllable start from their positions at stop closure”.
Based on the discussion of edge synchronization thus far, the movement toward the vowel in a CV syllable starts neither from the voice onset after the consonant release, nor from inside the closure, but from the onset of the final formant transition in the preceding syllable, as illustrated by the dotted curves in (B) of Figure 12b. Even if the syllable is utterance initial, as in the case of Delattre et al. [166], vowel target approximation also would have begun before the consonant closure, i.e., at the time when the consonant closure starts to form. Hence, the onset of the visible formant transition is well after the onset of vowel articulation. This perspective has two implications. Firstly, the linearity of the locus equations is largely due to a part-whole correlation [172,173,174], since the two F2 measurements are taken from two locations along the same unidirectional movement from the consonant to the vowel: voice onset, which is virtually a halfway point of the C-V transition, and center of the vowel, which is the end of the vowel target approximation. Secondly, because the vowel is coproduced with the consonant at syllable onset, and coarticulation resistance is the result of coproduction (depicted in (B) in Figure 12b as the warping of the dotted curves) whose severity depends on how much the consonant articulation conflicts with that of the vowel, to the extent resolvable by DSSTA, the slope of the locus equations would naturally reflect the amount of coarticulation resistance. DSSTA, therefore, has offered an ultimate solution to the inevitable conflict between consonants and vowels when they have to be coproduced, as per demand of CV synchronization.

3.2.4. Synchronization of Laryngeal and Supralaryngeal Articulations

As early as 1984, Ohala and Kawasaki suggested that “the division of sound sequences into syllables” is “for the sake of synchronizing the segmental and suprasegmental articulations” [175] (p. 123). This proposal foreshadowed a series of later findings in both tone and non-tone languages that generated five lines of evidence for the full synchrony of tone and syllables. The first line of evidence is that the start of tonal movement is aligned to syllable onset. The clearest cases are from tone languages, for which it is possible to directly observe where different tones start to move toward their respective targets [5,93,94,95]. In Figure 5, for example, in each plot, the F0 contours of the four tones start to diverge roughly from the onset of the second syllable toward their respective targets. A similar consistent start of F0 movement toward tonal targets at syllable onset has also been reported for Cantonese [140,144] and Shanghai Chinese [176]. For non-tone languages, there have been many findings of systematic alignment of F0 turning point to syllable onset, e.g., Dutch [177,178], Spanish [179], Greek [180], English [181,182], Italian [183], Portuguese [184], German [185], Arabic [186], and Persian [187].
The second line of evidence is that tonal target approximation starts from syllable onset even if the initial consonant is voiceless. As found in Mandarin [188], Cantonese [189], and English [190], when F0 contours are time-normalized with respect to the syllable, they parallel each other closely whether the onset consonant is sonorant as in [ma] or voiceless as in [ta], [tha], or [ʂa]. Thus, the approximation of the underlying tonal target starts not from the voice onset, but from the beginning of the syllable, regardless of whether the vocal folds are vibrating during the initial consonant.
The third line of evidence is that tonal target approximation ends at syllable offset even if there is a coda consonant. As found in [93], in disyllabic words with or without a nasal coda in the first syllable, the final part of tonal target approximation is executed through the coda nasal as if it is part of the first syllable. This indicates that the entire syllable is the domain of tonal target approximation whether or not it has a nasal coda, as long as the next syllable does not start with an approximant or a vowel.
The fourth line of evidence is that synchronization of laryngeal and supralaryngeal articulation may benefit vocal learning by eliminating temporal degrees of freedom [191]. As simulated with the qTA model, the learning of tones from raw f0 contours in real-speech data yields both faster and more efficient tone learning when tonal targets are assumed to be fully synchronized with the syllable than when the tone–syllable alignment is assumed to be flexible and has to be learned.
Finally, if C and V both start their target approximation earlier than the conventional syllable boundary as has been argued, then so should lexical tones. Initial evidence for the early tone onset can be seen in Figure 5 where the H tone in syllable 3 is followed by the L tone. F0 starts to drop toward the low target not at the conventional syllable boundary where the vertical line is drawn, but well before it (but also see discussion in Section 3.1.4). More solid evidence has been shown in [192], which applied the method developed in [153]. The results show that tone and vowel onsets are fully synchronized, which in turn indicates full synchrony of tone and syllable. Most surprisingly, with the newly established tone onset, the ‘anticipatory raising’ effect of tone is found to occur within rather than before the articulatory syllable. What this suggests is that the preparatory move discussed in Section 3.1.4 may be of two kinds, as can be seen in Figure 8. A slower one is from frame 1 to frame 3, which involves a backward extension of the arm and shoulder, and a fast one between frames 3 and 4, which involves mainly a backward flexion of the wrist. The anticipatory raising effect of the Low tone seems to be more like the fast preparatory wrist flexion, given its briefness. The finding that it happens within the syllable is astounding, and may have implications for many nonspeech motor movements as well.

3.2.5. Vowel Harmony, an Unresolved Issue

Vowel harmony is the phenomenon that in some languages, within a relatively large temporal domain such as a word or phrase, there is a tendency for vowels to share a particular property along a phonetic dimension, such as tongue height, or tongue frontness [87,193]. Such long-distance assimilation can be either left to right or right to left. Related to the right-to-left harmony is the long-distance anticipatory vowel-to-vowel assimilation across more than one syllable reported for English, which is referred to as a form of coarticulation [194,195]. Both kinds of long-distance assimilation seem incompatible with either target approximation or TD/AP, as it would mean that the approximation of a single vowel target can occur across multiple target approximation movements of the intervening consonants and vowels. The target approximation model, however, allows a distinction between target assignment and target approximation [196]. That is, it is possible for the target of a phone, be it segmental or suprasegmental, to be modified, or reassigned, before the start of its articulatory execution. The target reassignment, if extensive enough to be heard as a different vowel, could have originated historically from surface assimilation [197], due to listeners’ misperception [175,198]. Or it could be a small target readjustment in anticipation of an upcoming vowel [193] or consonant [199,200]. The separation of target assignment and target approximation as different processes means that only the latter involves genuine coarticulation, which is universal, while the former is due to a separate mechanism that is language specific, as has been demonstrated in a preliminary study of French [201].

3.2.6. Summary and Implications of Edge Synchronization

Multiple lines of evidence have been presented that consonants, vowels, and tones are likely synchronized by their onset at the beginning of a syllable. There is also a synchronization tendency at the offset of the syllable, although the evidence is only in terms of tone–syllable alignment. More discussion of syllable offset will occur in light of tactile anchoring in the next Section.
A major impact of the edge synchronization hypothesis is an overhaul to the way we conceptually segment speech, as illustrated in Figure 13. Unlike the conventional segmentation in Figure 13A, in Figure 13B, the onset of each segment is at a time when the spectrogram starts to move toward its prototypical configuration. The first /a/, for example, is in the middle of the conventional /i/ where F2 starts to drop, and the second /i/ is in the middle of the conventional /a/ where F2 starts to rise. Thus, the onset of a vowel is fully aligned to the onset of the initial consonant, which is also shifted leftward: to where an oral closure just starts to form, as indicated by the downward turn of F1. As explained earlier, the large leftward shift of a vowel onset (by about 100 ms) is a conceptual change that can explain away much of the anticipatory V-to-V coarticulation. The new offset of a segment is where the spectral pattern has maximally approached its canonical configuration and started to move away from it. For /i/, it is at the peak of F2 and F3, for /a/, it is at the peak of F1 and the valley of F2, and for /w/, it is at the valley of F2. For the obstruent consonants, the offset is not at the end of its prototypical spectral pattern (e.g., closure gap in /b/, nasal or lateral formants in /m/ and /l/, and the frication in /ʂ/), but in the middle of these intervals. Furthermore, a coda consonant, e.g., /n/ in /shan/, is aligned after the nuclear vowel, without overlap. The reason for the lack of VC overlap will be discussed in Section 3.3.2. Finally, tones are fully synchronized with the entire CV or CVC syllable, as shown in the bottom tier.
Just as importantly, Figure 13B makes it clear that the coproduction of consonants and vowels is in fact acoustically transparent rather than hidden. For example, the rise of F2 and F3 toward the high extremes of /i/ in /li/ from the middle of the conventional /a/ is clearly visible. Also apparent is the start of the F1 drop toward the low extreme in /l/ from the middle of the conventional /a/, which marks the onset of oral closure for both /l/ and /i/. Currently, these visible patterns are given names like transitions, anticipatory movements, etc. The new segmentation treats them, instead, as belonging to the main bodies of the segments.
The representation of the true segmental intervals by the new segmentation shown in Figure 13B is in fact foreshadowed by the concept of diphone in speech technology [203]. A diphone is defined as an acoustic chunk consisting of two adjacent halves of a pair of conventionally segmented phones [202,203]. As illustrated in Figure 13C, each diphone extends from the middle of one conventional phone to the middle of the next. Intriguingly, the diphone boundaries in Figure 13C match well with those of the new segmental intervals in Figure 13B. This means that a diphone actually represents a single phone rather than two phones in many cases, especially in the case of consonants. For example, the diphone [im], [al], [iw], and [eish] in Figure 13C actually represent the full scope of [m], [l], [w], and [sh], respectively. But the diphone representation of vowels is incomplete, compared to the new segmentation scheme, because it misses the initial portion of the vowel. For example, the vowel interval in [ma] in Figure 13B spans across two diphones in Figure 13C: [im] and [ma], and the vowel interval in [li] spans across [al] and [li]. Missing the initial portion of the vowel in the diphone segmentation therefore is likely a major reason why so many contextual features are needed in training a diphone synthesis or recognition system.

3.3. Tactile Anchoring

Tactile anchoring is about how synchronization is achieved in speech production, and it may hold the key to understanding some of the structural details about the syllable. The hypothesis is that the accuracy of edge synchronization rests on sensory feedback, and that tactile sensations generated during articulation likely provide the most useful feedback information. It follows that the points of synchronization are at the syllable edges rather than in the center of the syllable. Most previous theories of the syllable regard the center, where sonority is the highest, as the core of the syllable (see detailed review in [204]). Tactile anchoring predicts, in contrast, that the center of the syllable, where contact sensation is likely weak, would be the least reliable anchor.

3.3.1. Why Is Tactile Anchoring Needed?

One of the earliest clues comes from the finding that, just like bimanual synchronization [24], concurrent leg swinging by two people sitting next to each other also shows stable phase relations only at 0° and 180°, and 0° is the only stable relation at high speed [27]. But this holds only when the participating subjects can see each other’s movements. This perceptual nature of motor synchrony is further demonstrated by Mechsner et al. [26], which shows that the propensity for, as well as the ability to achieve bimanual synchrony are perceptual in nature. They demonstrate that naïve subjects are able to perform bimanual oscillations in a 4:3 frequency ratio, which are virtually impossible to maintain based purely on body-oriented strategies, as long as they can see a 1:1 frequency ratio converted from their actions by a mechanical device. Besides visual perception, tactile [205,206,207,208] and proprioceptive [209,210,211,212,213] information has also been shown to help stabilize in-phase coordination in bimanual tasks. Thus, the perceptual guidance needed for motor synchrony includes any sensory feedback, and the importance of each perceptual channel is a function of the clarity of the feedback information it provides to the central control system.
For speech, to ensure synchrony in syllable articulation, visual feedback is unlikely to be useful, as speakers cannot see their own articulators. Auditory feedback is available all the time and is likely very useful [214], but it may not be the most critical, as people who become deaf post-lingually are often able to speak intelligibly for decades [215,216]. Also abundantly available is proprioceptive feedback during speech, but the information it provides is likely spread evenly over time; thus, it may be useful, but it is not the most critical. The sensory information that probably fluctuates the most with the opening and closing of the vocal tract is from tactile feedback, especially from the articulators that are rich in tactile receptors, such as tongue tip, tongue blade, and the lips [217,218], whose sensitivity “rivals or exceeds that of the fingertip” [219]. This points to consonant closures as the most likely sites of tactile anchoring, because they are brief, easily palpable, and exact in time.

3.3.2. Evidence for Tactile Anchoring in Speech

Tactile anchoring is the most speculative component of the new syllable theory as there is only indirect evidence so far. The first is that blocking tactile feedback in the oral cavity through topical anesthesia not only lowers intelligibility, but also reduces speech rate [220], presumably because more time is needed to ensure tactile anchoring. The second is that in consonants that involve multiple articulatory components, the gestural components with a tighter oral contact tend to be aligned closer to syllable edges. In English, for example, the apical gesture in /l/ reaches its extreme near the syllable margin, whereas its dorsal component reaches the extreme closer to the nuclear vowel, whether /l/ is a coda (hence, the dark variant) or an onset (hence, the light variant) [221]. This means that gestures that generate clearer tactile feedback are preferred at syllable edges over those that provide less clear tactile information. That is, the apical gesture of /l/ involves a tongue tip contact with the alveolar ridge; hence, the rich tactile sensation at the tongue tip provides much more sensory feedback than the more vowel-like tongue body gesture [218]. A similar finding is that in /w/, the labial gesture is also more peripheral than the tongue body gesture [222]. Not only does the labial gesture of /w/ involve more skin contact than the tongue body gesture, but also the lips have a rich sensory representation [218].
The second phenomenon is the well-known onset-coda asymmetry; i.e., CV syllables are much more common than VC and CVC syllables, both within and across languages [47,62,69,75,175,223,224]. Even if they are already present in a language, coda consonants are more vulnerable than onset consonants, as they are subject to reduction, deletion, and resyllabification [61,79,225,226]. The vulnerability of the coda means that it is not as reliable as the onset for providing tactile feedback. As for why the onset/coda asymmetry is there in the first place, there are a number of possible reasons. First, syllable onset is where the greatest number of syllabic components can be synchronized, including consonant, vowel, and tone, as mentioned before. In contrast, syllable offset can end with either a vowel or a coda consonant, but not both. This is because the closure of a coda consonant is in direct conflict with the opening movement of the preceding vowel. This differs from the syllable onset, where it is tolerable for a vowel to be briefly interrupted by the closing movement of the initial consonant. The sequential articulation of codas is likely one of the major reasons for their vulnerability to reduction and deletion [114]. As a syllable shortens when the speech rate increases, there is less and less time left to allow sequential execution of multiple segments within the syllable [113]. This vulnerability means that syllable onset is the only temporal location for generating reliable tactile input. Furthermore, because target approximation is frequently incomplete [115], and different syllabic components may have different degrees of incomplete approximation, synchronizing their offsets is hard. Most importantly, syllable offset is also the onset of the next syllable, which already provides a synchronization point. So, there is no need for a coda to perform synchronization except at the end of an utterance.
The onset-coda asymmetry is also reflected in the resyllabification phenomenon, whereby a coda consonant goes through a change that makes it sound like the onset of the next syllable. This may happen within a word, e.g., ending, producing, which becomes en–ding, pro–du–cing, or across words, e.g., let us, thin air, which become le–tus, thi–nair. In language teaching, such resyllabification (often referred to as linking) is considered a good marker of fluency for languages like English, as non-native speakers often fail to do it [227,228]. There is doubt, however, as to whether resyllabification actually occurs, especially across word boundaries [35]. One of the reasons is that studies of resyllabification have generated diverse findings based on the researcher’s own intuition, native listener’s judgment [77,78,79,229,230], phonotactic analysis [70,71,76], or language-specific phonetic properties [35,231].
In Liu and Xu [232], we used a more objective method to determine syllable affiliation of intervocalic nasals at word boundaries in Southern British English. We used both singleton and cluster consonants spoken at a slow speech rate as benchmarks for canonical onset and coda consonants, and then used deep learning and dynamic time warping (DTW) to determine if some of the codas at a normal speech rate are classified as onset consonants. The majority of codas at a normal rate were indeed identified as onsets by the slow-speech-trained classifier, and the resyllabified codas were acoustically similar to their canonical onset counterparts. Also, the resyllabified coda consonants contained the same amount of information for the vowel of the second syllable as the canonical onset consonants, indicating that the resyllabified consonants were indeed coarticulated with the following vowel.
The propensity for resyllabification is further seen in a phenomenon first observed by Stetson [56]. He found that a CVC sequence such as pup, pup, pup…, when spoken at an increasing speech rate, changes abruptly at some point to a CV sequence: pu, pu, pu…. Kelso et al. [28], in a more formal experiment, show that a VC sequence like ip, ip, ip… also changes abruptly to pi, pi, pi … when the speaking rate is increased to about four syllables/s. A similar finding was reached by de Jong [233]. An abrupt shift at four syllables/s is striking, since normal speech rate is much faster, at about 5–7 syllables/s [6,234]. This means that resyllabification is virtually inevitable in normal speech, as is found in [232], unless codas are deleted when the following syllable starts with a vowel, as is the case in Mandarin [226].

3.3.3. Summary of Tactile Anchoring

The need for tactile anchoring is evident from the finding that the quality of bimanual synchrony of cyclic movements is contingent on the quality of perceptual guidance during the execution of a synchronization task ([26] and many others cited above). Assuming that motor synchrony is the essence of the syllable as currently hypothesized, its accuracy would require clear feedback guidance. Of all the sensory channels available during speech production, the intermittently fluctuating tactile feedback from consonants provides the most precise feedback. Given the vulnerability of codas, the only temporal location for tactile feedback is syllable onset. This hypothesis is supported by the fact that in sonorant consonants /l/ and /w/, the tongue tip and labial gestures that generate rich tactile information are realized near the syllable edges, while the tongue body gestures are realized toward the center of the syllable [221,222], and by the onset-coda asymmetry that strongly favors CV over VC or CVC. Further support in terms of the onset-coda asymmetry comes from the strong tendency for resyllabification of coda consonants to the onset of the next syllable [28,56,232,233]. One thing that is unclear, however, is how frequently tactical anchoring is needed. The many cases of vowel hiatus across syllable boundaries suggest that it is not mandatory for every syllable, but the strong propensity for resyllabification in languages like English and Dutch suggests that it is likely to be as frequent as possible. But this is still an open question for future research.

3.4. What Is New Compared to Previous Theories

As mentioned in Section 2, most of the well-known theories of the syllable, including the maximal onset principle [71,72,73], the sonority theories [71,235,236,237], the phonotactic theories [46,47,70], are about syllabification; i.e., how to divide continuous texts into separate syllables. None of them is particularly concerned with how syllables are articulated—the main focus of the current hypothesis. Theories that do consider the articulation of syllables, e.g., the chest pulse theory [56], the C/D model [57], and the frame/content theory [58], are not concerned about the DOF problem. Even the coupled-oscillator model [13] and the time structure model of the syllable [114], which both proposed CV synchronization, did not discuss the problem of degrees of freedom. Thus, the current hypothesis is the only theory, to our knowledge, that posits the syllable as a synchronization mechanism for solving the DOF problem to make human speech possible.
The synchronization principle may nevertheless be relevant to the issue of syllabification, however, as it suggests a novel way of identifying syllable boundaries that are based on articulatory timing rather than phonotactics. For example, it is shown that the coda consonant in the first syllable in a CVC#VC sequence in English is resyllabified into the following VC syllable based on evidence consistent with the synchronization principle [232]. And it is also shown that in an English CCV syllable, V is articulatorily aligned with the first C [157], which suggests the possibility that even in a C1V1C2# C3V2C4 sequence, the coda C2 of the first syllable may also become aligned with V2, thus forming a C2C3 cluster that is likely non-canonical in a language. But this has yet to be empirically tested.
The synchronization hypothesis also differs from other models of the syllable in that it is able to address issues beyond syllabification. Through the proposal of the DSSTA mechanism that resolves the conflict between consonant and vocalic articulation, coherent explanations can be offered for phenomena like coarticulation resistance, locus, and locus equation. And the inclusion of laryngeal gestures in the synchronization model enables the hypothesis to be connected to research on tone and intonation, thus fulfilling the prophecy of Ohala and Kawasaki [175].

4. Neural Prerequisites for Syllable Articulation

The discussion so far has presented arguments from the perspective of motor control regarding why the syllable is likely a synchronization mechanism for reducing degrees of freedom to make speech possible. No neural evidence has been presented, however, because there are no neural findings, to our knowledge, that would directly implicate a synchronization mechanism. More importantly, we are of the view that speech-related neural activities serve the purpose of making speech communication possible, rather than the latter being the byproduct of the former. The three core mechanisms of the synchronization hypothesis, nevertheless, may suggest specific neural substrates that are needed for the articulation of the syllable.
Target approximation, as discussed in Section 3 (cf., in particular, the graphic illustration in Figure 4 and Figure 6), implies that the neural commands sent to the articulatory muscles are in the form of underlying targets rather than either surface displacement or velocity [238,239] trajectories. Only in this way can the contextual variability in surface acoustic trajectories due to physical laws, mainly inertia, be articulatorily generated. Target approximation points out a clear forward relation from articulation to acoustics, which can be learned through analysis by synthesis [240]. In this process, the articulatory system repeatedly generates surface trajectories until a best fit is found, and the articulatory maneuver that can generate the best fit is stored as the learned target, as has been computationally simulated [111,119,163]. Analysis by synthesis may require that continuous acoustic signals remain available during learning to serve as auditory templates for imitative learning [241,242]. A recent set of computational simulations of speech acquisition showed that, surprisingly, the most effective way of learning the underlying target of consonants, vowels in English [117,120,121] and tones in Mandarin [243], is under the guidance of trained speech recognizers that simulate speech perception. This suggests that, while articulatory targets may indeed require separate neural representations, the auditory representation needed for guiding the acquisition of the articulatory targets is likely only implicit in the perceptual neural network rather than being localized in the brain.
Note that these latest findings have shown evidence that vocal learning is unlikely driven by online feedback control as is assumed in some of the most influential neural models, such as DIVA and HSFC [244]. Rather, vocal learning is more likely a process of prolonged trial-and-error target discovery, guided by phonological perception, with little or no immediate sensory feedback correction. The often-observed auditory feedback correction (due to the ease of laboratory elicitation) is likely an entirely separate neural process with the sole function of calibrating the articulatory-to-acoustic relation, just like the tuning of a musical instrument. Target approximation, in contrast, is hard to be immediately corrected not only because of the slow reaction time [245], but also because target undershoot happens all the time as a function of syllable duration [115]. The motor control of target approximation, therefore, likely relies entirely on a feedforward neural mechanism [238].
For edge synchronization, two critical neural mechanisms need to be in place. Firstly, there should be centrally generated signals for initiating each syllable. These signals, probably in the form of neural pulses, cannot be periodic, however, because their intervals need to frequently change with linguistic and paralinguistic factors such as stress [246], position in word and phrase [247,248], prosodic focus [94,233], and speaking style [249]. This view is consistent with findings of neural-based research on music and language [250]. Secondly, there should be neural mechanisms to coactivate all the involved articulators without significant time delay or discrepancy. One way to achieve this is to bring the neural control areas close to each other in the brain to ensure rapid communication. This may indeed have happened during the evolution of the human brain. Belyk [251] (p. 180) suggests that an evolutionary reorganization has brought expiration, phonation, and articulation into proximity in the brain, creating a small-world architecture [252,253] that would function efficiently. It would be interesting to examine in future research whether and how exactly this small-world architecture enables the synchronization of laryngeal, supralaryngeal, consonantal, and vocalic articulations.
The likely neural prerequisite for edge synchronization has been suggested by a recent neural theory of the evolutionary basis of beat-based dancing behavior shared by only humans and parrots [254]. The theory notes the human–parrot similarities in both movement to music and the neurobiology of advanced vocal learning, and suggests that “gene regulation changes associated with the evolution of a dorsal laryngeal pitch control pathway in ancestral humans fortuitously strengthened auditory–parietal cortical connections that support beat-based rhythmic processing” (p. 1). Hickok, in a commentary on Patel’s theory [255], suggests further that the emergence of the ability to synchronize to beats is likely a byproduct of the ability of speech to coordinate the timing of laryngeal pitch control and the supralaryngeal movements [256], which is exactly like what is proposed in the synchronization model of the syllable, short of only CV synchrony (echoing also the Ohala and Kawasaki prophecy [175]). Further work is therefore needed to reveal the neural mechanisms that underlie edge synchronization. One likely avenue is through technologies developed for treating neural disorders like epilepsy, including ECoG and SEEG [257,258], which may allow the observation of neural activities at sufficiently high time resolution corresponding to CVT synchrony.
For tactile anchoring, as discussed in Section 3.3, the precision of synchronization depends on the quality of sensorimotor feedback [26]. There should therefore be sensorimotor pathways that enable effective feedback control. The critical role of timing control has been recently demonstrated in songbirds by [259,260], showing that disorders like stuttering can be induced in Zebra finch by modifying the gene critical for timing control. Interestingly, the induced changes did not affect the structure of individual syllables in the bird songs. This is consistent with the synchronization hypothesis in which tactile anchoring and target approximation are two separate mechanisms, the former relying crucially on feedback control, while the latter relying mainly on feedforward control [102,238,261].
Even with the right genetic disposition, not only the ability to control the key articulators, but also the pathway to the brainstem may need time to fully develop after birth before synchronization can be attempted. This could be why canonical babbling, and with it, the ability to produce syllables, starts to emerge only around 6 months after birth [262], and even the order of the appearance of lingual consonants involved in the babbling follows that of the development of the tactile receptors in the tongue [217].
In summary, the synchronization hypothesis proposed in this paper may encourage future studies to look for specific neural activities that correspond to target approximation, edge synchronization, and tactile anchoring, or to demonstrate their implausibility. Given that the hypothesis is about the control of timing in motor movements, its neural correlates should show sufficient sensitivity to the temporal aspects of speech production. There has been some progress in this direction [257,263]. Future studies, however, need to be more purposefully designed to generate sufficiently specific evidence to either confirm or reject the synchronization hypothesis.

5. Conclusions and Broader Implications

This paper has proposed a new hypothesis about the syllable, positing it as a synchronization mechanism that makes the central nervous control of multiple articulatory movements in speech production possible by eliminating most of the temporal degrees of freedom. The hypothesis postulates three specific mechanisms: target approximation, edge synchronization, and tactile anchoring, which work together to execute the syllable—a recurring synchronization of multiple articulatory movements.
The proposed model differs from previous models of the syllable in that it is highly explicit, which makes it easily falsifiable. It can be rejected or at least weakened if it is demonstrated that reducing degrees of freedom is unnecessary, reducing temporal degrees of freedom via synchronization does not adequately resolve the DOF problem or simply does not happen, the target approximation model is defective, DSSTA is implausible, or tactile feedback is absent or ineffective.
Since its initial proposal in the form of the time–structure model of the syllable [114], various aspects of the synchronization model have received empirical support. Edge synchronization is supported by articulatory and acoustic evidence for pitch–syllable synchrony in Mandarin [188] and English [190], consonant–vowel synchrony in Mandarin and English [153,155,156,157], vowel–tone synchrony in Mandarin [192], and resyllabification in English [232]. The benefit of reducing degrees of freedom by edge synchronization as well as dimension-specific sequential target approximation is supported by computational simulation of tone learning [191] and the learning of English words [121]. The synchronization mechanism has therefore already received strong support, and its benefit has also been demonstrated by vocal learning simulations. The establishment of this synchronization mechanism, especially with the elaboration of the dimension-specific sequential target approximation (DSSTA), cf. Section 3.1.2 and Section 3.2.2, offers a coherent account of many unresolved phenomena of speech, including formant transitions, coarticulation, coarticulation resistance, undershoot and assimilation, locus and locus equation, resyllabification, asymmetry of syllable onset and coda, pre-low raising of pitch, etc. Finally, the importance of tactile feedback is shown by the adverse effect of topical oral anesthetics on intelligibility and speech rate [220].
The synchronization hypothesis may have a number of broader implications. (A) It suggests that it is likely that the emergence of syllable articulation makes it possible for human vocalization to be particulated into discrete consonants, vowels, and tones—units capable of making phonological contrasts, which can then be combined into words, phrases, and sentences. Given that chimpanzees and bonobos, our closest relatives, cannot be taught to speak, syllable articulation could be one of the most critical steps in human language evolution. (B) The recent modeling simulation of vocal learning [121] suggests that canonical babbling in the form of onset-synchronized syllable sequences that starts typically from 7 months old could be a prerequisite for vocal learning [264]. (C) The synchronization hypothesis, itself partially motivated by the problem of too many degrees of freedom [3], may have implications for motor control in general [265,266,267], as well as for robotics [267], as any complex motor movements, either natural or bionic, would have to face exactly the same problem. And (D) the hypothesis may also have implications for the understanding and diagnoses of speech disorders, especially those related to articulation. Different types of disorders may be due to difficulty with either target approximation, edge synchronization, or tactile anchoring.

Funding

This work was supported in part by the National Institutes of Health (NIH) Grant No. 1R01DC03902 and the Leverhulme Trust Research Project Grant RPG-2019-241.

Institutional Review Board Statement

Ethical review and approval were not applicable for this study as it did not involve human or animal subjects.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

Acknowledgments

I would like to thank Fang Liu, Santitham Prom-on, Hong Gao, Anqi Xu, Peter Birkholz, Daniel van Niekerk, Paul Krug, Zirui Liu, Yue Chen, Hanyi Meng, and Weiyi Kang for their contributions to the development of this work. The author, however, is responsible for all the errors.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Ladefoged, P. A Course in Phonetics; Hartcourt Brace Jovanovich: New York, NY, USA, 1982. [Google Scholar]
  2. Huys, R. The Dynamical Organization of Limb Movements. In Nonlinear Dynamics in Human Behavior; Springer: Berlin/Heidelberg, Germany, 2010; pp. 69–90. [Google Scholar]
  3. Bernstein, N.A. The Co-Ordination and Regulation of Movements; Pergamon Press: Oxford, UK, 1967. [Google Scholar]
  4. Stevens, K.N. Acoustic Phonetics; The MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  5. Xu, Y. Contextual tonal variations in Mandarin. J. Phon. 1997, 25, 61–83. [Google Scholar] [CrossRef]
  6. Tiffany, W.R. The effects of syllable structure on diadochokinetic and reading rates. J. Speech Hear. Res. 1980, 23, 894–908. [Google Scholar] [CrossRef] [PubMed]
  7. Latash, M.L.; Scholz, J.P.; Schöner, G. Toward a new theory of motor synergies. Mot. Control. 2007, 11, 276–308. [Google Scholar] [CrossRef] [PubMed]
  8. Easton, T.A. On the normal use of reflexes: The hypothesis that reflexes form the basic language of the motor program permits simple, flexible specifications of voluntary movements and allows fruitful speculation. Am. Sci. 1972, 60, 591–599. [Google Scholar]
  9. Turvey, M.T. Preliminaries to a theory of action with reference to vision. In Perceiving, Acting, and Knowing: Toward an Ecological Psychology; Shaw, R., Bransford, J., Eds.; Lawrence Erlbaum: Hillsdale, NJ, USA, 1977; pp. 211–265. [Google Scholar]
  10. Fowler, C.A.; Rubin, P.; Remez, R.E.; Turvey, M.T. Implications for speech production of a general theory of action. In Language Production; Butterworth; Academic Press: New York, NY, USA, 1980; pp. 373–420. [Google Scholar]
  11. Saltzman, E.L.; Munhall, K.G. A dynamical approach to gestural patterning in speech production. Ecol. Psychol. 1989, 1, 333–382. [Google Scholar] [CrossRef]
  12. Browman, C.P.; Goldstein, L. Articulatory phonology: An overview. Phonetica 1992, 49, 155–180. [Google Scholar] [CrossRef]
  13. Goldstein, L.; Byrd, D.; Saltzman, E. The role of vocal tract gestural action units in understanding the evolution of phonology. In Action to Language via the Mirror Neuron System; Arbib, M.A., Ed.; Cambridge University Press: Cambridge, UK, 2006; pp. 215–249. [Google Scholar]
  14. Nam, H.; Goldstein, L.; Saltzman, E. Self-organization of syllable structure: A coupled oscillator model. In Approaches to Phonological Complexity; Pellegrino, F., Marsico, E., Chitoran, I., Coupé, C., Eds.; Mouton de Gruyter: New York, NY, USA, 2009; pp. 299–328. [Google Scholar]
  15. Saltzman, E.; Nam, H.; Krivokapić, J.; Goldstein, L. A task-dynamic toolkit for modeling the effects of prosodic structure on articulation. Proc. Speech Prosody 2008, 2008, 175–184. [Google Scholar]
  16. Haken, H.; Kelso, J.A.S.; Bunz, H. A Theoretical Model of Phase Transitions in Human Hand Movements. Biol. Cybern. 1985, 51, 347–356. [Google Scholar] [CrossRef]
  17. Kay, B.; Kelso, J.; Saltzman, E.; Schöner, G. Space–time behavior of single and bimanual rhythmical movements: Data and limit cycle model. J. Exp. Psychol. Hum. Percept. Perform. 1987, 13, 178. [Google Scholar] [CrossRef]
  18. Semjen, A.; Ivry, R.B. The coupled oscillator model of between-hand coordination in alternate-hand tapping: A reappraisal. J. Exp. Psychol. Hum. Percept. Perform. 2001, 27, 251. [Google Scholar] [CrossRef]
  19. Huygens, C. A letter to his father, dated 26 Feb. 1665. In Ouevres Completes de Christian Huyghens; Nijhoff, M., Ed.; Societe Hollandaise des Sciences: The Hague, The Netherlands, 1893; Volume 5, p. 243. [Google Scholar]
  20. Pikovsky, A.; Rosenblum, M.; Kurths, J. Synchronization—A Universal Concept in Nonlinear Sciences; Cambridge University Press: Cambridge, UK, 2001. [Google Scholar]
  21. Kelso, J.A.S.; Southard, D.L.; Goodman, D. On the nature of human interlimb coordination. Science 1979, 203, 1029–1031. [Google Scholar] [CrossRef] [PubMed]
  22. Cummins, F.; Li, C.; Wang, B. Coupling among speakers during synchronous speaking in English and Mandarin. J. Phon. 2013, 41, 432–441. [Google Scholar] [CrossRef]
  23. Cummins, F. Periodic and Aperiodic Synchronization in Skilled Action. Front. Hum. Neurosci. 2011, 5, 170. [Google Scholar] [CrossRef] [PubMed]
  24. Kelso, J.A.S. Phase transitions and critical behavior in human bimanual coordination. Am. J. Physiol. Regul. Integr. Comp. 1984, 246, R1000–R1004. [Google Scholar] [CrossRef]
  25. Kelso, J.A.S.; Tuller, B.; Harris, K.S. A “dynamic pattern” perspective on the control and coordination of movement. In The Production of Speech; MacNeilage, P.F., Ed.; Springer-Verlag: New York, NY, USA, 1983; pp. 137–173. [Google Scholar]
  26. Mechsner, F.; Kerzel, D.; Knoblich, G.; Prinz, W. Perceptual basis of bimanual coordination. Nature 2001, 414, 69–73. [Google Scholar] [CrossRef]
  27. Schmidt, R.C.; Carello, C.; Turvey, M.T. Phase transitions and critical fluctuations in the visual coordination of rhythmic movements between people. J. Exp. Psychol. Hum. Percept. Perform. 1990, 16, 227–247. [Google Scholar] [CrossRef]
  28. Kelso, J.A.S.; Saltzman, E.L.; Tuller, B. The dynamical perspective on speech production: Data and theory. J. Phon. 1986, 14, 29–59. [Google Scholar] [CrossRef]
  29. Adler, R. A study of locking phenomena in oscillators. Proc. IRE 1946, 34, 351–357. [Google Scholar] [CrossRef]
  30. Bennett, M.; Schatz, M.F.; Rockwood, H.; Wiesenfeld, K. Huygens’s clocks. Proc. Math. Phys. Eng. Sci. 2002, 458, 563–579. [Google Scholar] [CrossRef]
  31. DeFrancis, J.F. Visible Speech: The Diverse Oneness of Writing Systems; University of Hawaii Press: Honolulu, HI, USA, 1989. [Google Scholar]
  32. Gnanadesikan, A.E. The Writing Revolution: Cuneiform to the Internet; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  33. Liberman, I.Y.; Shankweiler, D.; Fischer, F.W.; Carter, B. Explicit syllable and phoneme segmentation in the young child. J. Exp. Child Psychol. 1974, 18, 201–212. [Google Scholar] [CrossRef]
  34. Fox, B.; Routh, D. Analyzing spoken language into words, syllables, and phonomes: A developmental study. J. Psycholinguist. Res. 1975, 4, 331–342. [Google Scholar] [CrossRef]
  35. Shattuck-Hufnagel, S. The role of the syllable in speech production in American English: A fresh consideration of the evidence. In Handbook of the Syllable; Cairns, C.E., Raimy, E., Eds.; Brill: Boston, MA, USA, 2011; pp. 197–224. [Google Scholar]
  36. Bolinger, D. Contrastive accent and contrastive stress. Language 1961, 37, 83–96. [Google Scholar] [CrossRef]
  37. de Jong, K. Stress, lexical focus, and segmental focus in English: Patterns of variation in vowel duration. J. Phon. 2004, 32, 493–516. [Google Scholar] [CrossRef]
  38. Pierrehumbert, J. The Phonology and Phonetics of English Intonation. Ph.D. Dissertation, Massachusetts Institutes of Technology, Cambridge, MA, USA, 1980. [Google Scholar]
  39. Barbosa, P.; Bailly, G. Characterisation of rhythmic patterns for text-to-speech synthesis. Speech Commun. 1994, 15, 127–137. [Google Scholar] [CrossRef]
  40. Cummins, F.; Port, R. Rhythmic constraints on stress timing in English. J. Phon. 1998, 26, 145–171. [Google Scholar] [CrossRef]
  41. Nolan, F.; Asu, E.L. The Pairwise Variability Index and Coexisting Rhythms in Language. Phonetica 2009, 66, 64–77. [Google Scholar] [CrossRef]
  42. Abramson, A.S. Static and dynamic acoustic cues in distinctive tones. Lang. Speech 1978, 21, 319–325. [Google Scholar] [CrossRef]
  43. Chao, Y.R. A Grammar of Spoken Chinese; University of California Press: Berkeley, CA, USA, 1968. [Google Scholar]
  44. Selkirk, E.O. On prosodic structure and its relation to syntactic structure. In Nordic Prosody II, Trondheim, Norway; Fretheim, T., Ed.; Indiana University Linguistics Club: Bloomington, IN, USA, 1978; pp. 111–140, TAPIR. [Google Scholar]
  45. Nespor, M.; Vogel, I. Prosodic Phonology; Foris Publications: Dordrecht, the Netherlands, 1986. [Google Scholar]
  46. Blevins, J. The syllable in phonological theory. In Handbook of Phonological Theory; Goldsmith, J., Ed.; Blackwell: Cambridge, MA, USA, 2001; pp. 206–244. [Google Scholar]
  47. Hooper, J.B. The syllable in phonological theory. Language 1972, 48, 525–540. [Google Scholar] [CrossRef]
  48. Bertoncini, J.; Mehler, J. Syllables as units in infant speech perception. Infant Behav. Dev. 1981, 4, 247–260. [Google Scholar] [CrossRef]
  49. Content, A.; Kearns, R.K.; Frauenfelder, U.H. Boundaries versus onsets in syllabic segmentation. J. Mem. Lang. 2001, 45, 177–199. [Google Scholar] [CrossRef]
  50. Cutler, A.; Mehler, J.; Norris, D.; Segui, J. The syllable’s differing role in the segmentation of French and English. J. Mem. Lang. 1986, 25, 385–400. [Google Scholar] [CrossRef]
  51. Kohler, K.J. Is the syllable a phonological universal? J. Linguist. 1966, 2, 207–208. [Google Scholar] [CrossRef]
  52. Labrune, L. Questioning the universality of the syllable: Evidence from Japanese. Phonology 2012, 29, 113–152. [Google Scholar] [CrossRef]
  53. Gimson, A.C. An Introduction to the Pronunciation of English; Arnold: London, UK, 1970. [Google Scholar]
  54. Steriade, D. Alternatives to syllable-based accounts of consonantal phonotactics. In Proceedings of the Linguistics and Phonetics 1998: Item Order in Language and Speech; Fujimura, O., Joseph, B.D., Palek, B., Eds.; Karolinum Press: Prague, Czech Republic, 1999; pp. 205–245. [Google Scholar]
  55. Blevins, J. Evolutionary Phonology: The Emergence of Sound Patterns; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  56. Stetson, R.H. Motor Phonetics: A study of Speech Movements in Action; North Holland: Amsterdam, The Netherlands, 1951. [Google Scholar]
  57. Fujimura, O. C/D Model: A computational model of phonetic implementation. In Language and Computations; Ristad, E.S., Ed.; American Math Society: Providence, RI, USA, 1994; pp. 1–20. [Google Scholar]
  58. MacNeilage, P.F. The frame/content theory of evolution of speech production. Behav. Brain Sci. 1998, 21, 499–546. [Google Scholar] [CrossRef]
  59. Fitch, W.T. The Evolution of Language; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar]
  60. Pinker, S. The Language Instinct: The New Science of Language and Mind; Penguin: London, UK, 1995. [Google Scholar]
  61. Dell, G.S. The retrieval of phonological forms in production: Tests of predictions from a connectionist model. J. Mem. Lang. 1988, 27, 124–142. [Google Scholar] [CrossRef]
  62. Levelt, W.J.M.; Roelofs, A.; Meyer, A.S. A theory of lexical access in speech production. Behav. Brain Sci. 1999, 22, 1–38. [Google Scholar] [CrossRef]
  63. Jakobson, R.; Fant, C.G.; Halle, M. Preliminaries to Speech Analysis. The Distinctive Features and Their Correlates; MIT Press: Cambridge, MA, USA, 1951. [Google Scholar]
  64. Turk, A.; Nakai, S.; Sugahara, M. Acoustic Segment Durations in Prosodic Research: A Practical Guide. In Methods in Empirical Prosody Research; Sudhoff, S., Lenertová, D., Meyeretal, R., Eds.; De Gruyter: Berlin/Heidelberg, Germany; New York, NY, USA, 2006; pp. 1–28. [Google Scholar]
  65. Peterson, G.E.; Lehiste, I. Duration of syllable nuclei in English. J. Acoust. Soc. Am. 1960, 32, 693–703. [Google Scholar] [CrossRef]
  66. Hockett, C.F. A Manual of Phonology (International Journal of American Linguistics, Memoir 11); Waverly Press: Baltimore, MD, USA, 1955. [Google Scholar]
  67. Farnetani, E.; Recasens, D. Coarticulation and connected speech processes. Handb. Phon. Sci. 1997, 371, 404. [Google Scholar]
  68. Kühnert, B.; Nolan, F. The origin of coarticulation. In Coarticulation: Theory, Data and Techniques; Hardcastle, W.J., Newlett, N., Eds.; Cambridge University Press: Cambridge, UK, 1999; pp. 7–30. [Google Scholar]
  69. Vennemann, T. Preference Laws for Syllable Structure and the Explanation of Sound Change; Mouton de Gruyter: Berlin, Germany, 1988. [Google Scholar]
  70. Pulgram, E. Syllable, Word, Nexus, Cursus; The Hague: Mouton, France, 1970. [Google Scholar]
  71. Steriade, D. Greek Prosodies and the Nature of Syllabification. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1982. [Google Scholar]
  72. Hoard, J.E. Aspiration, tenseness, and syllabication in English. Language 1971, 47, 133–140. [Google Scholar] [CrossRef]
  73. Wells, J.C. Syllabification and allophony. In Studies in the Pronunciation of English: A Commemorative Volume in Honour of A. C. Gimson; Ramsaran, S., Ed.; Routledge: London, UK, 1990; pp. 76–86. [Google Scholar]
  74. Fudge, E. Syllables. J. Linguist. 1969, 5, 253–286. [Google Scholar] [CrossRef]
  75. Selkirk, E.O. The syllable. In The Structure of Phonological Representations, Part II; Hulst, H.V.D., Smith, N., Eds.; Foris: Dordrecht, The Netherlands, 1982; pp. 337–383. [Google Scholar]
  76. Duanmu, S. Syllable Structure: The Limits of Variation; Oxford University Press: Oxford, UK, 2009. [Google Scholar]
  77. Chiosáin, M.N.; Welby, P.; Espesser, R. Is the syllabification of Irish a typological exception? An experimental study. Speech Commun. 2012, 54, 68–91. [Google Scholar] [CrossRef]
  78. Goslin, J.; Frauenfelder, U.H. A Comparison of Theoretical and Human Syllabification. Lang. Speech 2001, 44, 409–436. [Google Scholar] [CrossRef] [PubMed]
  79. Schiller, N.O.; Meyer, A.S.; Levelt, W.J. The syllabic structure of spoken words: Evidence from the syllabification of intervocalic consonants. Lang. Speech 1997, 40, 103–140. [Google Scholar] [CrossRef] [PubMed]
  80. Menzerath, P.; de Lacerda, A. Koartikulation, Seuerung und Lautabgrenzung; Dummlers: Berlin/Bonn, Germany, 1933. [Google Scholar]
  81. Kozhevnikov, V.A.; Chistovich, L.A. Speech: Articulation and Perception; Translation by Joint Publications Research Service: Washington, DC, USA, 1965; JPRS 30543. [Google Scholar]
  82. Gay, T. Articulatory units: Segments or syllables. In Syllables and Segments; Bell, A., Hooper, J., Eds.; North-Holland: Amsterdam, The Netherlands, 1978; pp. 121–132. [Google Scholar]
  83. Kent, R.; Minifie, F. Coarticulation in recent speech production models. J. Phon. 1977, 5, 115–133. [Google Scholar] [CrossRef]
  84. Kent, R.D.; Moll, K.L. Tongue Body Articulation during Vowel and Diphthong Gestures. Folia Phoniatr. Logop. 1972, 24, 278–300. [Google Scholar] [CrossRef] [PubMed]
  85. Moll, K.; Daniloff, R. Investigation of the timing of velar movement during speech. J. Acoust. Soc. Am. 1971, 50, 678–684. [Google Scholar] [CrossRef] [PubMed]
  86. Öhman, S.E.G. Coarticulation in VCV utterances: Spectrographic measurements. J. Acoust. Soc. Am. 1966, 39, 151–168. [Google Scholar] [CrossRef] [PubMed]
  87. Clements, G.N. Vowel Harmony in Nonlinear Generative Phonology; Indiana University Linguistics Club: Bloomington, IN, USA, 1976. [Google Scholar]
  88. Huffman, M.K. Measures of phonation type in Hmong. J. Acoust. Soc. Am. 1987, 81, 495–504. [Google Scholar] [CrossRef]
  89. Wayland, R.; Jongman, A. Acoustic correlates of breathy and clear vowels: The case of Khmer. J. Phon. 2003, 31, 181–201. [Google Scholar] [CrossRef]
  90. Lindblom, B. Spectrographic study of vowel reduction. J. Acoust. Soc. Am. 1963, 35, 1773–1781. [Google Scholar] [CrossRef]
  91. Fujisaki, H. Dynamic characteristics of voice fundamental frequency in speech and singing. In The Production of Speech; MacNeilage, P.F., Ed.; Springer-Verlag: New York, NY, USA, 1983; pp. 39–55. [Google Scholar]
  92. Xu, Y.; Wang, Q.E. Pitch targets and their realization: Evidence from Mandarin Chinese. Speech Commun. 2001, 33, 319–337. [Google Scholar] [CrossRef]
  93. Xu, Y. Consistency of tone-syllable alignment across different syllable structures and speaking rates. Phonetica 1998, 55, 179–203. [Google Scholar] [CrossRef] [PubMed]
  94. Xu, Y. Effects of tone and focus on the formation and alignment of F0 contours. J. Phon. 1999, 27, 55–105. [Google Scholar] [CrossRef]
  95. Xu, Y. Fundamental frequency peak delay in Mandarin. Phonetica 2001, 58, 26–52. [Google Scholar] [CrossRef]
  96. Bailly, G.; Holm, B. SFC: A trainable prosodic model. Speech Commun. 2005, 46, 348–364. [Google Scholar] [CrossRef]
  97. van Santen, J.; Kain, A.; Klabbers, E.; Mishra, T. Synthesis of prosody using multi-level unit sequences. Speech Commun. 2005, 46, 365–375. [Google Scholar] [CrossRef]
  98. Browman, C.P.; Goldstein, L. Articulatory gestures as phonological units. Phonology 1989, 6, 201–251. [Google Scholar] [CrossRef]
  99. Byrd, D.; Saltzman, E. The elastic phrase: Modeling the dynamics of boundary-adjacent lengthening. J. Phon. 2003, 31, 149–180. [Google Scholar] [CrossRef]
  100. Fowler, C.A. Coarticulation and theories of extrinsic timing. J. Phon. 1980, 8, 113–133. [Google Scholar] [CrossRef]
  101. Xu, Y. Timing and coordination in tone and intonation—An articulatory-functional perspective. Lingua 2009, 119, 906–927. [Google Scholar] [CrossRef]
  102. Prom-on, S.; Xu, Y.; Thipakorn, B. Modeling tone and intonation in Mandarin and English as a process of target approximation. J. Acoust. Soc. Am. 2009, 125, 405–424. [Google Scholar] [CrossRef]
  103. Alzaidi, M.S.A.; Xu, Y.; Xu, A.; Szreder, M. Analysis and computational modelling of Emirati Arabic intonation—A preliminary study. J. Phon. 2023, 98, 101236. [Google Scholar] [CrossRef]
  104. Lee, A.; Simard, C.; Tamata, A.; Xu, Y.; Prom-on, S.; Sun, J. Modelling Fijian focus prosody using PENTAtrainer: A pilot study. In Proceedings of the 2nd International Conference on Tone and Intonation (TAI 2023), Singapore, 18–21 November 2023; pp. 9–10. [Google Scholar]
  105. Lee, A.; Xu, Y. Modelling Japanese intonation using pentatrainer2. In Proceedings of the 18th International Congress of Phonetic Sciences, Glasgow, UK, 10–14 August 2015; International Phonetic Association: Glasgow, UK, 2015. [Google Scholar]
  106. Ouyang, I.C. Non-segmental cues for syllable perception: The role of local tonal f0 and global speech rate in syllabification. IJCLCLP 2013, 18, 59–79. [Google Scholar]
  107. Simard, C.; Wegener, C.; Lee, A.; Chiu, F.; Youngberg, C. Savosavo word stress: A quantitative analysis. In Proceedings of the Speech Prosody, Dublin, Ireland, 20–23 May 2014; pp. 512–514. [Google Scholar]
  108. Ta, T.Y.; Ngo, H.H.; Van Nguyen, H. A New Computational Method for Determining Parameters Representing Fundamental Frequency Contours of Speech Words. J. Inf. Hiding Multim. Signal Process. 2020, 11, 1. [Google Scholar]
  109. Taheri-Ardali, M.; Xu, Y. An articulatory-functional approach to modeling Persian focus prosody. In Proceedings of the 18th International Congress of Phonetic Sciences, Glasgow, UK, 10–14 August 2015; pp. 326–329. [Google Scholar]
  110. Thai, T.Y.; Huy, H.N.; Tuyet, D.V.; Ablameyko, S.V.; Hung, N.V.; Hoa, D.V. Tonal languages speech synthesis using an indirect pitch markers and the quantitative target approximation methods. J. Belarusian State Univ. Math. Inform. 2019, 3, 105–121. [Google Scholar] [CrossRef]
  111. Xu, Y.; Prom-on, S. Toward invariant functional representations of variable surface fundamental frequency contours: Synthesizing speech melody via model-based stochastic learning. Speech Commun. 2014, 57, 181–208. [Google Scholar] [CrossRef]
  112. Liu, F.; Xu, Y.; Prom-on, S.; Yu, A.C.L. Morpheme-like prosodic functions: Evidence from acoustic analysis and computational modeling. J. Speech Sci. 2013, 3, 85–140. [Google Scholar] [CrossRef]
  113. Cheng, C.; Xu, Y. Articulatory limit and extreme segmental reduction in Taiwan Mandarin. J. Acoust. Soc. Am. 2013, 134, 4481–4495. [Google Scholar] [CrossRef]
  114. Xu, Y.; Liu, F. Tonal alignment, syllable structure and coarticulation: Toward an integrated model. Ital. J. Linguist. 2006, 18, 125–159. [Google Scholar]
  115. Xu, Y.; Prom-on, S. Economy of Effort or Maximum Rate of Information? Exploring Basic Principles of Articulatory Dynamics. Front. Psychol. 2019, 10, 2469. [Google Scholar] [CrossRef]
  116. Birkholz, P.; Kroger, B.J.; Neuschaefer-Rube, C. Model-Based Reproduction of Articulatory Trajectories for Consonant-Vowel Sequences. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 1422–1433. [Google Scholar] [CrossRef]
  117. Krug, P.K.; Birkholz, P.; Gerazov, B.; Niekerk, D.R.V.; Xu, A.; Xu, Y. Artificial vocal learning guided by phoneme recognition and visual information. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 1734–1744. [Google Scholar] [CrossRef]
  118. Prom-on, S.; Birkholz, P.; Xu, Y. Training an articulatory synthesizer with continuous acoustic data. Proc. Interspeech 2013, 2013, 349–353. [Google Scholar]
  119. Prom-on, S.; Birkholz, P.; Xu, Y. Identifying underlying articulatory targets of Thai vowels from acoustic data based on an analysis-by-synthesis approach. EURASIP J. Audio Speech Music. Process. 2014, 2014, 23. [Google Scholar] [CrossRef]
  120. Van Niekerk, D.R.; Xu, A.; Gerazov, B.; Krug, P.K.; Birkholz, P.; Halliday, L.; Prom-on, S.; Xu, Y. Simulating vocal learning of spoken language: Beyond imitation. Speech Commun. 2023, 147, 51–62. [Google Scholar] [CrossRef]
  121. Xu, A.; van Niekerk, D.R.; Gerazov, B.; Krug, P.K.; Birkholz, P.; Prom-on, S.; Halliday, L.F.; Xu, Y. Artificial vocal learning guided by speech recognition: What it may tell us about how children learn to speak. J. Phon. 2024, 105, 101338. [Google Scholar] [CrossRef]
  122. Moon, S.-J.; Lindblom, B. Interaction between duration, context, and speaking style in English stressed vowels. J. Acoust. Soc. Am. 1994, 96, 40–55. [Google Scholar] [CrossRef]
  123. Browman, C.P.; Goldstein, L. Targetless schwa: An articulatory analysis. In Papers in Laboratory Phonology II: Gesture, Segment, Prosody; Docherty, G.J., Ladd, R., Eds.; Cambridge University Press: Cambridge, UK, 1992; pp. 26–36. [Google Scholar]
  124. Wood, S.A.J. Assimilation or coarticulation? Evidence from the temporal co-ordination of tongue gestures for the palatalization of Bulgarian alveolar stops. J. Phon. 1996, 24, 139–164. [Google Scholar] [CrossRef]
  125. Bell-Berti, F.; Krakow, R.A. Anticipatory velar lowering: A coproduction account. J. Acoust. Soc. Am. 1991, 90, 112–123. [Google Scholar] [CrossRef]
  126. Boyce, S.E.; Krakow, R.A.; Bell-Berti, F. Phonological underspecification and speech motor organization. Phonology 1991, 8, 210–236. [Google Scholar] [CrossRef]
  127. Chen, Y.; Xu, Y. Production of weak elements in speech -- Evidence from f0 patterns of neutral tone in standard Chinese. Phonetica 2006, 63, 47–75. [Google Scholar] [CrossRef] [PubMed]
  128. Ostry, D.J.; Gribble, P.L.; Gracco, V.L. Coarticulation of jaw movements in speech production: Is context sensitivity in speech kinematics centrally planned? J. Neurosicence 1996, 16, 1570–1979. [Google Scholar] [CrossRef] [PubMed]
  129. Laboissiere, R.; Ostry, D.J.; Feldman, A.G. The control of multi-muscle systems: Human jaw and hyoid movements. Biol. Cybern. 1996, 74, 373–384. [Google Scholar] [CrossRef] [PubMed]
  130. Byrd, D.; Kaun, A.; Narayanan, S.; Saltzman, E. Phrasal signatures in articulation. In Papers in Laboratory Phonology V: Acquisition and the Lexicon; Broe, M.B., Pierrehumbert, J.B., Eds.; Cambridge University Press: Cambridge, UK, 2000; pp. 70–87. [Google Scholar]
  131. Edwards, J.R.; Beckman, M.E.; Fletcher, J. The articulatory kinematics of final lengthening. J. Acoust. Soc. Am. 1991, 89, 369–382. [Google Scholar] [CrossRef]
  132. Arvaniti, A.; Ladd, D.R. Underspecification in intonation revisited: A reply to Xu, Lee, Prom-on and Liu. Phonology 2015, 32, 537–541. [Google Scholar] [CrossRef]
  133. Keating, P.A. Underspecification in phonetics. Phonology 1988, 5, 275–292. [Google Scholar] [CrossRef]
  134. Myers, S. Surface underspecification of tone in Chichewa. Phonology 1998, 15, 367–392. [Google Scholar] [CrossRef]
  135. Steriade, D. Underspecification and markedness. In Handbook of Phonological Theory; Goldsmith, J.A., Ed.; Basil Blackweell: Oxford, UK, 1995; pp. 114–174. [Google Scholar]
  136. Fujisaki, H.; Wang, C.; Ohno, S.; Gu, W. Analysis and synthesis of fundamental frequency contours of Standard Chinese using the command–response model. Speech Commun. 2005, 47, 59–70. [Google Scholar] [CrossRef]
  137. Whalen, D.H. Coarticulation is largely planned. J. Phon. 1990, 18, 3–35. [Google Scholar] [CrossRef]
  138. Saitou, T.; Unoki, M.; Akagi, M. Development of an F0 control model based on F0 dynamic characteristics for singing-voice synthesis. Speech Commun. 2005, 46, 405–417. [Google Scholar] [CrossRef]
  139. Gandour, J.; Potisuk, S.; Dechongkit, S. Tonal coarticulation in Thai. J. Phon. 1994, 22, 477–492. [Google Scholar] [CrossRef]
  140. Gu, W.; Lee, T. Effects of tonal context and focus on Cantonese F0. In Proceedings of the 16th International Congress of Phonetic Sciences, Saarbrucken, Germany, 6–10 August 2007; pp. 1033–1036. [Google Scholar]
  141. Laniran, Y.O.; Clements, G.N. Downstep and high raising: Interacting factors in Yoruba tone production. J. Phon. 2003, 31, 203–250. [Google Scholar] [CrossRef]
  142. Lee, A.; Xu, Y.; Prom-On, S. Pre-low raising in Japanese pitch accent. Phonetica 2017, 74, 231–246. [Google Scholar] [PubMed]
  143. hubpages.com. How to Hit a Great Smash in Badminton. 2014. Available online: http://hubpages.com/games-hobbies/Badminton-Smash-How-to-Play-the-Shot# (accessed on 8 September 2016).
  144. Wong, Y.W. Contextual Tonal Variations and Pitch Targets in Cantonese. In Proceedings of the Speech Prosody 2006, Dresden, Germany, 2–5 May 2006. PS3-13-199. [Google Scholar]
  145. Gay, T. Effect of speaking rate on diphthong formant movements. J. Acoust. Soc. Am. 1968, 44, 1570–1573. [Google Scholar] [CrossRef]
  146. Fowler, C.A. Production and perception of coarticulation among stressed and unstressed vowels. J. Speech Hear. Res. 1981, 46, 127–139. [Google Scholar]
  147. Brunner, J.; Geng, C.; Sotiropoulou, S.; Gafos, A. Timing of German onset and word boundary clusters. Lab. Phonol. 2014, 5, 403–454. [Google Scholar]
  148. Gao, M. Gestural Coordination among vowel, consonant and tone gestures in Mandarin Chinese. Chin. J. Phon. 2009, 2, 43–50. [Google Scholar]
  149. Shaw, J.A.; Chen, W.-R. Spatially-conditioned speech timing: Evidence and implications. Front. Psychol. 2019, 10, 2726. [Google Scholar] [CrossRef]
  150. Šimko, J.; O’Dell, M.; Vainio, M. Emergent consonantal quantity contrast and context-dependence of gestural phasing. J. Phon. 2014, 44, 130–151. [Google Scholar]
  151. Yi, H.; Tilsen, S. Interaction between lexical tone and intonation: An EMA study. Proc. Interspeech 2016, 2016, 2448–2452. [Google Scholar]
  152. Tilsen, S. Detecting anticipatory information in speech with signal chopping. J. Phon. 2020, 82, 100996. [Google Scholar] [CrossRef]
  153. Liu, Z.; Xu, Y.; Hsieh, F.-f. Coarticulation as synchronised CV co-onset – Parallel evidence from articulation and acoustics. J. Phon. 2022, 90, 101116. [Google Scholar] [CrossRef]
  154. Xu, Y. Speech as articulatory encoding of communicative functions. In Proceedings of the 16th International Congress of Phonetic Sciences, Saarbrucken, Germany, 6–10 August 2007; pp. 25–30. [Google Scholar]
  155. Xu, Y.; Gao, H. FormantPro as a tool for speech analysis and segmentation. Rev. Estud. Ling. 2018, 26, 1435–1454. [Google Scholar] [CrossRef]
  156. Xu, Y.; Liu, F. Determining the temporal interval of segments with the help of F0 contours. J. Phon. 2007, 35, 398–420. [Google Scholar] [CrossRef]
  157. Liu, Z.; Xu, Y. Segmental alignment of English syllables with singleton and cluster onsets. In Proceedings of the Interspeech 2021, Brno, Czech Republic, 30 August–3 September 2021. [Google Scholar]
  158. Browman, C.P.; Goldstein, L.M. Competing constraints on intergestural coordination and self-organization of phonological structures. Cah. l’ICP. Bull. Commun. Parlée 2000, 5, 25–34. [Google Scholar]
  159. Bladon, R.A.W.; Al-Bamerni, A. Coarticulation resistance of English /l/. J. Phon. 1976, 4, 135–150. [Google Scholar] [CrossRef]
  160. Recasens, D. Vowel-to-vowel coarticulation in Catalan VCV sequences. J. Acoust. Soc. Am. 1984, 76, 1624–1635. [Google Scholar] [CrossRef]
  161. Recasens, D. V-to-C coarticulation in Catalan VCV sequences: An articulatory and acoustical study. J. Phon. 1984, 12, 61–73. [Google Scholar] [CrossRef]
  162. Dembowski, J.; Lindstrom, M.J.; Westbury, J.R. Articulator point variability in the production of stop consonants. In Neuromotor Speech Disorders: Nature, Assessment, and Management; Cannito, M.P., Yorkston, K.M., Beukelman, D.R., Eds.; Paul H. Brookes: Baltimore, MD, USA, 1998; pp. 27–46. [Google Scholar]
  163. Xu, A.; Birkholz, P.; Xu, Y. Coarticulation as synchronized dimension-specific sequential target approximation: An articulatory synthesis simulation. In Proceedings of the 19th International Congress of Phonetic Sciences, Melbourne, Australia, 5–9 August 2019. [Google Scholar]
  164. Birkholz, P.; Jackel, D. A three-dimensional model of the vocal tract for speech synthesis. In Proceedings of the 15th International Congress of Phonetic Sciences, Barcelona, Spain, 3–9 August 2003; pp. 2597–2600. [Google Scholar]
  165. Cooper, F.S.; Delattre, P.C.; Liberman, A.M.; Borst, J.M.; Gerstman, L.J. Some experiments on the perception of synthetic speech sounds. J. Acoust. Soc. Am. 1952, 24, 597–606. [Google Scholar] [CrossRef]
  166. Delattre, P.C.; Liberman, A.M.; Cooper, F.S. Acoustic Loci and Transitional Cues for Consonants. J. Acoust. Soc. Am. 1955, 27, 769–773. [Google Scholar] [CrossRef]
  167. Liberman, A.M.; Cooper, F.S.; Shankweiler, D.P.; Studdert-Kennedy, M.G. Perception of the speech code. Psychol. Rev. 1967, 74, 431–461. [Google Scholar] [CrossRef] [PubMed]
  168. Cooper, F.S.; Liberman, A.M.; Borst, J.M. The interconversion of audible and visible patterns as a basis for research in the perception of speech. Proc. Natl. Acad. Sci. USA 1951, 37, 318–325. [Google Scholar] [CrossRef] [PubMed]
  169. Lindblom, B.; Sussman, H.M. Dissecting coarticulation: How locus equations happen. J. Phon. 2012, 40, 1–19. [Google Scholar] [CrossRef]
  170. Fowler, C.A. Invariants, specifiers, cues: An investigation of locus equations as information for place of articulation. Percept. Psychophys. 1994, 55, 597–610. [Google Scholar] [CrossRef]
  171. Iskarous, K.; Fowler, C.A.; Whalen, D.H. Locus equations are an acoustic expression of articulator synergy. J. Acoust. Soc. Am. 2010, 128, 2021–2032. [Google Scholar] [CrossRef]
  172. Benoit, C. Note on the use of correlation in speech timing. J. Acoust. Soc. Am. 1986, 80, 1846–1849. [Google Scholar] [CrossRef]
  173. Löfqvist, A. Proportional timing in speech motor control. J. Phon. 1991, 19, 343–350. [Google Scholar] [CrossRef]
  174. Munhall, K.G. An examination of intra-articulator relative timing. J. Acoust. Soc. Am. 1985, 78, 1548–1553. [Google Scholar] [CrossRef]
  175. Ohala, J.J.; Kawasaki, H. Prosodic phonology and phonetics. Phonology 1984, 1, 113–127. [Google Scholar] [CrossRef]
  176. Ling, B.; Liang, J. Tonal alignment in shanghai Chinese. In Proceedings of the COCOSDA2015, Shanghai, China, 28–30 October 2015; pp. 128–132. [Google Scholar]
  177. Caspers, J.; van Heuven, V.J. Effects of time pressure on the phonetic realization of the Dutch accent-lending pitch rise and fall. Phonetica 1993, 50, 161–171. [Google Scholar] [CrossRef]
  178. Ladd, D.R.; Mennen, I.; Schepman, A. Phonological conditioning of peak alignment in rising pitch accents in Dutch. J. Acoust. Soc. Am. 2000, 107, 2685–2696. [Google Scholar] [CrossRef] [PubMed]
  179. Prieto, P.; Torreira, F. The segmental anchoring hypothesis revisited: Syllable structure and speech rate effects on peak timing in Spanish. J. Phon. 2007, 35, 473–500. [Google Scholar] [CrossRef]
  180. Arvaniti, A.; Ladd, D.R.; Mennen, I. Stability of tonal alignment: The case of Greek prenuclear accents. J. Phon. 1998, 36, 3–25. [Google Scholar] [CrossRef]
  181. Ladd, D.R.; Faulkner, D.; Faulkner, H.; Schepman, A. Constant “segmental anchoring” of F0 movements under changes in speech rate. J. Acoust. Soc. Am. 1999, 106, 1543–1554. [Google Scholar] [CrossRef] [PubMed]
  182. Xu, Y.; Xu, C.X. Phonetic realization of focus in English declarative intonation. J. Phon. 2005, 33, 159–197. [Google Scholar] [CrossRef]
  183. D’Imperio, M. Focus and tonal structure in Neapolitan Italian. Speech Commun. 2001, 33, 339–356. [Google Scholar] [CrossRef]
  184. Frota, S. Tonal association and target alignment in European Portuguese nuclear falls. In Laboratory Phonology VII; Gussenhoven, C., Warner, N., Eds.; Mouton de Gruyter: Berlin/Heidelberg, Germany, 2002; pp. 387–418. [Google Scholar]
  185. Atterer, M.; Ladd, D.R. On the phonetics and phonology of "segmental anchoring" of F0: Evidence from German. J. Phon. 2004, 32, 177–197. [Google Scholar] [CrossRef]
  186. Yeou, M. Effects of focus, position and syllable structure on F0 alignment patterns in Arabic. In Proceedings of the JEP-TALN 2004, Arabic Language Processing, Fez, Morocco, 19–22 April 2004; pp. 19–22. [Google Scholar]
  187. Sadat-Tehrani, N. The alignment of L+H* pitch accents in Persian intonation. J. Int. Phon. Assoc. 2009, 39, 205–230. [Google Scholar] [CrossRef]
  188. Xu, C.X.; Xu, Y. Effects of consonant aspiration on Mandarin tones. J. Int. Phon. Assoc. 2003, 33, 165–181. [Google Scholar] [CrossRef]
  189. Wong, Y.W.; Xu, Y. Consonantal perturbation of f0 contours of Cantonese tones. In Proceedings of the 16th International Congress of Phonetic Sciences, Saarbrucken, Germany, 6–10 August 2007; pp. 1293–1296. [Google Scholar]
  190. Xu, Y.; Xu, A. Consonantal F0 perturbation in American English involves multiple mechanisms. J. Acoust. Soc. Am. 2021, 149, 2877–2895. [Google Scholar] [CrossRef]
  191. Xu, Y.; Prom-on, S. Degrees of freedom in prosody modeling. In Speech Prosody in Speech Synthesis—Modeling, Realizing, Converting Prosody for High Quality and Flexible speech Synthesis; Hirose, K., Tao, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2015; pp. 19–34. [Google Scholar]
  192. Kang, W.; Xu, Y. Tone-syllable synchrony in Mandarin: New evidence and implications. Speech Commun. 2024, 163, 103121. [Google Scholar] [CrossRef]
  193. Nguyen, N.; Fagyal, Z. Acoustic aspects of vowel harmony in French. J. Phon. 2008, 36, 1–27. [Google Scholar] [CrossRef]
  194. Magen, H.S. The extent of vowel-to-vowel coarticulation in English. J. Phon. 1997, 25, 187–205. [Google Scholar] [CrossRef]
  195. Grosvald, M. Long-Distance Coarticulation in Spoken and Signed Language: An Overview. Lang. Linguist. Compass 2010, 4, 348–362. [Google Scholar] [CrossRef]
  196. Xu, Y. Speech melody as articulatorily implemented communicative functions. Speech Commun. 2005, 46, 220–251. [Google Scholar] [CrossRef]
  197. Gafos, A.I.; Benus, S. Dynamics of phonological cognition. Cogn. Sci. 2006, 30, 905–943. [Google Scholar] [CrossRef]
  198. Ohala, J.J. Towards a universal, phonetically-based, theory of vowel harmony. In Proceedings of the Third International Conference on Spoken Language Processing, Yokohama, Japan, 18–22 September 1994; pp. 491–494. [Google Scholar]
  199. Heid, S.; Hawkins, S. An acoustical study of long-domain/r/and/l/coarticulation. In Proceedings of the 5th Seminar on Speech Production: Models and Data, Kloster Seeon, Germany, 1–4 May 2000; pp. 77–80. [Google Scholar]
  200. West, P. The extent of coarticulation of English liquids: An acoustic and articulatory study. In Proceedings of the 14th International Congress of Phonetic Sciences, San Francisco, CA, USA, 1–7 August 1999; pp. 1901–1904. [Google Scholar]
  201. Chiu, F.; Fromont, L.; Lee, A.; Xu, Y. Long-distance anticipatory vowel-to-vowel assimilatory effects in French and Japanese. In Proceedings of the 18th International Congress of Phonetic Sciences, Glasgow, UK, 10–14 August 2015; pp. 1008–1012. [Google Scholar]
  202. Peterson, G.E.; Wang WS, Y.; Sivertsen, E. Segmentation techniques in speech synthesis. J. Acoust. Soc. Am. 1958, 30, 739–742. [Google Scholar] [CrossRef]
  203. Taylor, P. Text-to-Speech Synthesis; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  204. Ohala, J.J. Alternatives to the sonority hierarchy for explaining segmental sequential constraints. In Papers from the Parasession on the Syllable; Chicago Linguistic Society: Chicago, IL, USA, 1992; pp. 319–338. [Google Scholar]
  205. Buchanan, J.J.; Ryu, Y.U. The interaction of tactile information and movement amplitude in a multijoint bimanual circle-tracing task: Phase transitions and loss of stability. Q. J. Exp. Psychol. Sect. A 2005, 58, 769–787. [Google Scholar] [CrossRef]
  206. Johansson, R.S.; Flanagan, J.R. Coding and use of tactile signals from the fingertips in object manipulation tasks. Nat. Rev. Neurosci. 2009, 10, 345. [Google Scholar] [CrossRef]
  207. Kelso SJ, A.; Fink, P.W.; DeLaplain, C.R.; Carson, R.G. Haptic information stabilizes and destabilizes coordination dynamics. Proc. R. Soc. Lond. B Biol. Sci. 2001, 268, 1207–1213. [Google Scholar] [CrossRef]
  208. Koh, K.; Kwon, H.J.; Yoon, B.C.; Cho, Y.; Shin, J.-H.; Hahn, J.-O.; Miller, R.H.; Kim, Y.H.; Shim, J.K. The role of tactile sensation in online and offline hierarchical control of multi-finger force synergy. Exp. Brain Res. 2015, 233, 2539–2548. [Google Scholar] [CrossRef] [PubMed]
  209. Baldissera, F.; Cavallari, P.; Marini, G.; Tassone, G. Differential control of in-phase and anti-phase coupling of rhythmic movements of ipsilateral hand and foot. Exp. Brain Res. 1991, 83, 375–380. [Google Scholar] [CrossRef] [PubMed]
  210. Mechsner, F.; Stenneken, P.; Cole, J.; Aschersleben, G.; Prinz, W. Bimanual circling in deafferented patients: Evidence for a role of visual forward models. J. Neuralphsiology 2007, 1, 259–282. [Google Scholar] [CrossRef] [PubMed]
  211. Ridderikhoff, A.; Peper, C.E.; Beek, P.J. Error correction in bimanual coordination benefits from bilateral muscle activity: Evidence from kinesthetic tracking. Exp. Brain Res. 2007, 181, 31–48. [Google Scholar] [CrossRef] [PubMed]
  212. Spencer, R.M.C.; Ivry, R.B.; Cattaert, D.; Semjen, A. Bimanual coordination during rhythmic movements in the absence of somatosensory feedback. J. Neuralphsiology 2005, 94, 2901–2910. [Google Scholar] [CrossRef]
  213. Wilson, A.D.; Bingham, G.P.; Craig, J.C. Proprioceptive perception of phase variability. J. Exp. Psychol. Hum. Percept. Perform. 2003, 29, 1179–1190. [Google Scholar] [CrossRef]
  214. Masapollo, M.; Nittrouer, S. Immediate auditory feedback regulates inter-articulator speech coordination in service to phonetic structure. J. Acoust. Soc. Am. 2024, 156, 1850–1861. [Google Scholar] [CrossRef]
  215. Cowie, R.I.; Douglas-Cowie, E. Speech production in profound postlingual deafness. In Hearing Science and Hearing Disorders; Elsevier: Amsterdam, The Netherlands, 1983; pp. 183–230. [Google Scholar]
  216. Lane, H.; Webster, J.W. Speech deterioration in postlingually deafened adults. J. Acoust. Soc. Am. 1991, 89, 859–866. [Google Scholar] [CrossRef]
  217. Lyubimova, Z.V.; Sisengalieva, G.Z.; Chulkova, N.Y.; Smykova, O.I.; Selin, S.V. Role of tactile receptor structures of the tongue in speech sound production of infants of the first year of life. Bull. Exp. Biol. Med. 1999, 127, 115–119. [Google Scholar] [CrossRef]
  218. Ringel, R.L.; Ewanowski, S.J. Oral Perception: 1. Two-Point Discrimination. J. Speech Lang. Hear. Res. 1965, 8, 389–398. [Google Scholar] [CrossRef]
  219. Kent, R.D. The Feel of Speech: Multisystem and Polymodal Somatosensation in Speech Production. J. Speech Lang. Hear. Res. 2024, 67, 1424–1460. [Google Scholar] [CrossRef] [PubMed]
  220. De Letter, M.; Criel, Y.; Lind, A.; Hartsuiker, R.; Santens, P. Articulation lost in space. The effects of local orobuccal anesthesia on articulation and intelligibility of phonemes. Brain Lang. 2020, 207, 104813. [Google Scholar] [CrossRef] [PubMed]
  221. Sproat, R.; Fujimura, O. Allophonic variation in English /l/ and its implications for phonetic implementation. J. Phon. 1993, 21, 291–311. [Google Scholar] [CrossRef]
  222. Gick, B. Articulatory correlates of ambisyllabicity in English glides and liquids. In Papers in Laboratory Phonology VI: Constraints on Phonetic Interpretation; Local, J., Ogden, R., Temple, R., Eds.; Cambridge University Press: Cambridge, UK, 2003; pp. 222–236. [Google Scholar]
  223. Clements, G.N.; Keyser, S.J. CV phonology. a generative theory of the syllable. Linguist. Inq. Monogr. Camb. Mass. 1983, 9, 1–191. [Google Scholar]
  224. Locke, J.L. Phonological Acquisition and Change; Academic Press: London, UK, 1983. [Google Scholar]
  225. Gao, H.; Xu, Y. Ambisyllabicity in English: How real is it? In Proceedings of the 9th Phonetics Conference of China (PCC2010), Tianjin, China, 28 May–1 June 2010. [Google Scholar]
  226. Xu, Y. Acoustic-phonetic characteristics of junctures in Mandarin Chinese. Zhongguo Yuwen [J. Chin. Linguist.] 1986, 4, 353–360. (In Chinese) [Google Scholar]
  227. Anderson-Hsieh, J.; Riney, T.; Koehler, K. Connected speech modifications in the English of Japanese ESL learners. Ideal 1994, 7, 31–52. [Google Scholar]
  228. Hieke, A.E. Linking as a marker of fluent speech. Lang. Speech 1984, 27, 343–354. [Google Scholar] [CrossRef]
  229. Gaskell, M.G.; Spinelli, E.; Meunier, F. Perception of resyllabification in French. Mem. Cognit. 2002, 30, 798–810. [Google Scholar] [CrossRef]
  230. Treiman, R.; Danis, C. Syllabification of intervocalic consonants. J. Mem. Lang. 1988, 27, 87–104. [Google Scholar] [CrossRef]
  231. Strycharczuk, P.; Kohlberger, M. Resyllabification reconsidered: On the durational properties of word-final /s/ in Spanish. Lab. Phonol. 2016, 7, 1–24. [Google Scholar]
  232. Liu, Z.; Xu, Y. Deep learning assessment of syllable affiliation of intervocalic consonants. J. Acoust. Soc. Am. 2023, 153, 848–866. [Google Scholar] [CrossRef] [PubMed]
  233. de Jong, K. Rate-induced resyllabification revisited. Lang. Speech 2001, 44, 197–216. [Google Scholar] [CrossRef] [PubMed]
  234. Eriksson, A. Aural/acoustic vs. automatic methods in forensic phonetic case work. In Forensic Speaker Recognition: Law Enforcement and Counter-Terrorism; Neustein, A., Patil, H.A., Eds.; Springer: New York, NY, USA, 2012; pp. 41–69. [Google Scholar]
  235. Jespersen, O. Fonetik: En Systematisk Fremstilling af Læren om Sproglyd; Det Schøbergse Forlag: København, Denmark, 1899. [Google Scholar]
  236. Whitney, W.D. The relation of vowel and consonant. J. Am. Orient. Soc. 1865, 8, 357–373. [Google Scholar]
  237. Clements, G.N. The Role of the Sonority Cycle in Core Syllabification. In Papers in Laboratory Phonology 1: Between the Grammar and Physics of Speech; Beckman, M., Ed.; Cambridge University Press: Cambridge, UK, 1990; pp. 283–333. [Google Scholar]
  238. Guenther, F.H.; Ghosh, S.S.; Tourville, J.A. Neural modeling and imaging of the cortical interactions underlying syllable production. Brain Lang. 2006, 96, 280–301. [Google Scholar] [CrossRef]
  239. Tourville, J.A.; Guenther, F.H. The DIVA model: A neural theory of speech acquisition and production. Lang. Cogn. Process. 2011, 26, 952–981. [Google Scholar] [CrossRef]
  240. Cohen, A.; Hart, J.T. On the anatomy of intonation. Lingua 1967, 19, 177–192. [Google Scholar] [CrossRef]
  241. Kuhl, P.K. On babies, birds, modules, and mechanisms: A comparative approach to the acquisition of vocal communication. In The Comparative Psychology of Audition: Perceiving Complex Sounds; Dooling, R.J., Hulse, S.H., Eds.; Erlbaum: Hillsdale, NJ, USA, 1989; pp. 379–419. [Google Scholar]
  242. Margoliash, D.; Schmidt, M.F. Sleep, off-line processing, and vocal learning. Brain Lang. 2010, 115, 45–58. [Google Scholar] [CrossRef]
  243. Meng, H.; Chen, Y.; Liu, Z.; Xu, Y. Mandarin tone production can be learned under perceptual guidance—A machine learning simulation. In Proceedings of the 20th International Congress of Phonetic Sciences, Prague, Czech Republic, 7–11 August 2023; pp. 2324–2328. [Google Scholar]
  244. Guenther, F.H.; Hickok, G. Neural models of motor speech control. In Neurobiology of Language; Hickok, G., Small, S.L., Eds.; Elsevier: San Diego, CA, USA, 2016; pp. 725–740. [Google Scholar]
  245. Xu, Y.; Larson, C.R.; Bauer, J.J.; Hain, T.C. Compensation for pitch-shifted auditory feedback during the production of Mandarin tone sequences. J. Acoust. Soc. Am. 2004, 116, 1168–1178. [Google Scholar] [CrossRef]
  246. Fry, D.B. Experiments in the perception of stress. Lang. Speech 1958, 1, 126–152. [Google Scholar] [CrossRef]
  247. Nakatani, L.H.; O’connor, K.D.; Aston, C.H. Prosodic aspects of American English speech rhythm. Phonetica 1981, 38, 84–106. [Google Scholar] [CrossRef]
  248. Wang, C.; Xu, Y.; Zhang, J. Functional timing or rhythmical timing, or both? A corpus study of English and Mandarin duration. Front. Psychol. 2023, 13, 869049. [Google Scholar] [CrossRef] [PubMed]
  249. Baker, R.E.; Bradlow, A.R. Variability in word duration as a function of probability, speech style, and prosody. Lang. Speech 2009, 52, 391–413. [Google Scholar] [CrossRef] [PubMed]
  250. Patel, A.D. Music, Language, and the Brain; Oxford University Press: New York, NY, USA, 2008. [Google Scholar]
  251. Belyk, M.; Brown, S. The origins of the vocal brain in humans. Neurosci. Biobehav. Rev. 2017, 77, 177–193. [Google Scholar] [CrossRef]
  252. Sporns, O. Small-world connectivity, motif composition, and complexity of fractal neuronal connections. Biosystems 2006, 85, 55–64. [Google Scholar] [CrossRef]
  253. Sporns, O.; Zwi, J.D. The small world of the cerebral cortex. Neuroinformatics 2004, 2, 145–162. [Google Scholar] [CrossRef]
  254. Patel, A.D. Beat-based dancing to music has evolutionary foundations in advanced vocal learning. BMC Neurosci. 2024, 25, 65. [Google Scholar] [CrossRef]
  255. Hickok, G. The “coordination conjecture” as an alternative to Patel’s fortuitous enhancement hypothesis for the relation between vocal learning and beat-based dancing. BMC Neurosci. 2024, 25, 59. [Google Scholar] [CrossRef]
  256. Hickok, G.; Venezia, J.; Teghipco, A. Beyond Broca: Neural architecture and evolution of a dual motor speech coordination system. Brain 2023, 146, 1775–1790. [Google Scholar] [CrossRef]
  257. Dichter, B.K.; Breshears, J.D.; Leonard, M.K.; Chang, E.F. The control of vocal pitch in human laryngeal motor cortex. Cell 2018, 174, 21–31.e9. [Google Scholar] [CrossRef]
  258. Duraivel, S.; Rahimpour, S.; Chiang, C.-H.; Trumpis, M.; Wang, C.; Barth, K.; Harward, S.C.; Lad, S.P.; Friedman, A.H.; Southwell, D.G. High-resolution neural recordings improve the accuracy of speech decoding. Nat. Commun. 2023, 14, 6938. [Google Scholar] [CrossRef]
  259. Kubikova, L.; Bosikova, E.; Cvikova, M.; Lukacova, K.; Scharff, C.; Jarvis, E.D. Basal ganglia function, stuttering, sequencing, and repair in adult songbirds. Sci. Rep. 2014, 4, 6590. [Google Scholar] [CrossRef] [PubMed]
  260. Tanakaa, M.; Alvarado, J.S.; Murugan, M.; Mooney, R. Focal expression of mutant huntingtin in the songbird basal ganglia disrupts cortico-basal ganglia networks and vocal sequences. Proc. Natl. Acad. Sci. USA 2016, 113, E1720–E1727. [Google Scholar] [CrossRef] [PubMed]
  261. Perkell, J.S. Movement goals and feedback and feedforward control mechanisms in speech production. J. Neurolinguist. 2012, 25, 382–407. [Google Scholar] [CrossRef]
  262. Kuhl, P.K.; Meltzoff, A.N. Infant vocalizations in response to speech: Vocal imitation and developmental change. J. Acoust. Soc. Am. 1996, 100, 2425–2438. [Google Scholar] [CrossRef]
  263. Chartier, J.; Anumanchipalli, G.K.; Johnson, K.; Chang, E.F. Encoding of articulatory kinematic trajectories in human speech sensorimotor cortex. Neuron 2018, 98, 1042–1054.e4. [Google Scholar] [CrossRef]
  264. ter Haar, S.M.; Fernandez, A.A.; Gratier, M.; Knörnschild, M.; Levelt, C.; Moore, R.K.; Vellema, M.; Wang, X.; Oller, D.K. Cross-species parallels in babbling: Animals and algorithms. Philos. Trans. R. Soc. B Biol. Sci. 2021, 376, 20200239. [Google Scholar] [CrossRef]
  265. Grimme, B.; Fuchs, S.; Perrier, P.; Schöner, G. Limb versus speech motor control: A conceptual review. Motor Control 2011, 15, 5–33. [Google Scholar] [CrossRef]
  266. Konczak, J.; Vander Velden, H.; Jaeger, L. Learning to play the violin: Motor control by freezing, not freeing degrees of freedom. J. Mot. Behav. 2009, 41, 243–252. [Google Scholar] [CrossRef]
  267. Morasso, P. A vexing question in motor control: The degrees of freedom problem. Front. Bioeng. Biotechnol. 2022, 9, 783501. [Google Scholar] [CrossRef]
Figure 1. Spectrogram of the Mandarin phrase “比麻黎偽善” /bǐ má lí wěi shàn/ [more hypocritical than Ma Li], with broad phonetic transcriptions and pitch tract (in blue). C, V, and T stand for consonant, vowel, and tone, respectively. The segmentation is based on well-known conventions [63,64]. The segmentation of /w/ is based on Peterson and Lehiste (1960) [65].
Figure 1. Spectrogram of the Mandarin phrase “比麻黎偽善” /bǐ má lí wěi shàn/ [more hypocritical than Ma Li], with broad phonetic transcriptions and pitch tract (in blue). C, V, and T stand for consonant, vowel, and tone, respectively. The segmentation is based on well-known conventions [63,64]. The segmentation of /w/ is based on Peterson and Lehiste (1960) [65].
Brainsci 15 00033 g001
Figure 2. Temporal organization of articulatory dimensions under the synchronization hypothesis. The tiers represent the four articulatory domains controlled by the central nervous system: C for conant, V for vowel, T for tone, and Ph for phonation. The dotted curves represent asymptotic articulatory approximation of underlying targets (target approximation). The full alignment of the onsets and offsets of the approximation movements represent edge synchronization facilitated by tactile anchoring.
Figure 2. Temporal organization of articulatory dimensions under the synchronization hypothesis. The tiers represent the four articulatory domains controlled by the central nervous system: C for conant, V for vowel, T for tone, and Ph for phonation. The dotted curves represent asymptotic articulatory approximation of underlying targets (target approximation). The full alignment of the onsets and offsets of the approximation movements represent edge synchronization facilitated by tactile anchoring.
Brainsci 15 00033 g002
Figure 3. Interdependence of the three mechanisms of the synchronization hypothesis. Target approximation defines the onsets and offsets of individual movements; edge synchronization aligns movement onsets and offsets (rather than acoustic landmarks); and tactile anchoring provides the sensory feedback that ensures the accuracy of synchronization.
Figure 3. Interdependence of the three mechanisms of the synchronization hypothesis. Target approximation defines the onsets and offsets of individual movements; edge synchronization aligns movement onsets and offsets (rather than acoustic landmarks); and tactile anchoring provides the sensory feedback that ensures the accuracy of synchronization.
Brainsci 15 00033 g003
Figure 4. The target approximation model. A schematic illustration of hypothetical phonetic targets (dashed lines) and their surface realization (solid curve). The three vertical lines represent the boundaries of the two consecutive target intervals. The level dashed line on the right represents a static target, and the oblique dashed line on the left represents a dynamic target. In both intervals, the targets are asymptotically approximated. Adapted from the original version for tone and intonation [92].
Figure 4. The target approximation model. A schematic illustration of hypothetical phonetic targets (dashed lines) and their surface realization (solid curve). The three vertical lines represent the boundaries of the two consecutive target intervals. The level dashed line on the right represents a static target, and the oblique dashed line on the left represents a dynamic target. In both intervals, the targets are asymptotically approximated. Adapted from the original version for tone and intonation [92].
Brainsci 15 00033 g004
Figure 5. Mean time-normalized F0 contours of Mandarin tones in 5-syllable sentences, where all syllables are in the form of nasal + vowel. In each plot, the tones of all the syllables remain constant except those of the 2nd syllable, which alternate from High (H) to Rising (R), Low (L), and Falling (F). The tone of the third syllable varies from H in (A), F in (B), to R in (C). Data from Xu [94].
Figure 5. Mean time-normalized F0 contours of Mandarin tones in 5-syllable sentences, where all syllables are in the form of nasal + vowel. In each plot, the tones of all the syllables remain constant except those of the 2nd syllable, which alternate from High (H) to Rising (R), Low (L), and Falling (F). The tone of the third syllable varies from H in (A), F in (B), to R in (C). Data from Xu [94].
Brainsci 15 00033 g005
Figure 6. Sequential and blended target approximation processes generated with the qTA model [102]. The units of both axes are arbitrary. In (A), the three target approximation movements are strictly sequential, and the vertical line is the boundary between the first two movements. In (B), the vertical reference remains at time 0.15, but the first movement is shortened by 0.05 units. All the movements remain sequential (which means that the “tail” of the first movement—the dotted continuation—is truncated). In (C), the first and second movements overlap with each other by 0.05 units. The overlap is implemented by inserting a blended target (dotted green horizontal line), which is the average of the first two targets.
Figure 6. Sequential and blended target approximation processes generated with the qTA model [102]. The units of both axes are arbitrary. In (A), the three target approximation movements are strictly sequential, and the vertical line is the boundary between the first two movements. In (B), the vertical reference remains at time 0.15, but the first movement is shortened by 0.05 units. All the movements remain sequential (which means that the “tail” of the first movement—the dotted continuation—is truncated). In (C), the first and second movements overlap with each other by 0.05 units. The overlap is implemented by inserting a blended target (dotted green horizontal line), which is the average of the first two targets.
Brainsci 15 00033 g006aBrainsci 15 00033 g006b
Figure 7. F0 contours of the neutral tone vs. full tones in Mandarin: (A) Four full tones followed by a Falling tone. (B) The same four tones followed by a sequence of neutral tones. (C) A sequence of neutral tones followed by either a Falling tone or a Low tone. Data from Chen and Xu [127]. These plots show evidence that the neutral tone has a fully specified underlying pitch target, which includes a specification for a weak articulatory strength.
Figure 7. F0 contours of the neutral tone vs. full tones in Mandarin: (A) Four full tones followed by a Falling tone. (B) The same four tones followed by a sequence of neutral tones. (C) A sequence of neutral tones followed by either a Falling tone or a Low tone. Data from Chen and Xu [127]. These plots show evidence that the neutral tone has a fully specified underlying pitch target, which includes a specification for a weak articulatory strength.
Brainsci 15 00033 g007
Figure 8. Decomposition of a badminton smash consisting of three phases. Frames 1–4 are the preparation phase, while frames 4–6 are the unidirectional approximation phase. The goal is not only to reach the position of the racket-shuttle contact, but also to achieve a high velocity at the point of contact. Frames 7–9 are the settling phase [143], (courtesy of Michael Hayes at HowTheyPlay.com).
Figure 8. Decomposition of a badminton smash consisting of three phases. Frames 1–4 are the preparation phase, while frames 4–6 are the unidirectional approximation phase. The goal is not only to reach the position of the racket-shuttle contact, but also to achieve a high velocity at the point of contact. Frames 7–9 are the settling phase [143], (courtesy of Michael Hayes at HowTheyPlay.com).
Brainsci 15 00033 g008
Figure 9. A reinterpretation of Öhman [86]. Reproduced with permission from the Acoustical Society of America, with illustrative modifications. The orange highlight tracks F2 movement toward the vowel target, while the cyan highlight tracks F1 movement toward consonant closure.
Figure 9. A reinterpretation of Öhman [86]. Reproduced with permission from the Acoustical Society of America, with illustrative modifications. The orange highlight tracks F2 movement toward the vowel target, while the cyan highlight tracks F1 movement toward consonant closure.
Brainsci 15 00033 g009
Figure 10. Example spectrograms of a C1V1#C2V2 minimal triplet from [155]. The three-syllable sequence at the top serves as a control for both sequences at the bottom: /li/ in (A) contrasts with /ji/ in (B) for C2: /l/ vs. /j/, and with /lu/ in (C) for V2: /i/ vs. /u/. The red arrows point to the common divergent points for both C and V.
Figure 10. Example spectrograms of a C1V1#C2V2 minimal triplet from [155]. The three-syllable sequence at the top serves as a control for both sequences at the bottom: /li/ in (A) contrasts with /ji/ in (B) for C2: /l/ vs. /j/, and with /lu/ in (C) for V2: /i/ vs. /u/. The red arrows point to the common divergent points for both C and V.
Brainsci 15 00033 g010
Figure 11. Mean F2 trajectories of the middle three syllables in the Mandarin sentence frame “bi __ wei shan” [more hypocritical than __ ], where the keyword is a disyllabic personal name. The three trajectories in each plot differ either in the initial consonant or the vowel, and the four plots differ only in the first syllable. The time of all trajectories is relative to the voice onset of /bi/ in the carrier. Both F2 and time are averaged across 8 repetitions by 3 male speakers. (Data from [155]).
Figure 11. Mean F2 trajectories of the middle three syllables in the Mandarin sentence frame “bi __ wei shan” [more hypocritical than __ ], where the keyword is a disyllabic personal name. The three trajectories in each plot differ either in the initial consonant or the vowel, and the four plots differ only in the first syllable. The time of all trajectories is relative to the voice onset of /bi/ in the carrier. Both F2 and time are averaged across 8 repetitions by 3 male speakers. (Data from [155]).
Brainsci 15 00033 g011
Figure 12. (a) An illustration of the locus phenomenon [166]. Reproduced with permission from the Acoustical Society of America). The curves are F1 (the curves near the bottom in both plots), and F2 hand-painted for the pattern playback speech synthesizer [168]. b, d, and g mark how listeners identified the consonants. (b) A reinterpretation of [166] with added virtual F2 traces in red. In A, F2 is fully continuous; in B, the initial portion of F2 is silenced.
Figure 12. (a) An illustration of the locus phenomenon [166]. Reproduced with permission from the Acoustical Society of America). The curves are F1 (the curves near the bottom in both plots), and F2 hand-painted for the pattern playback speech synthesizer [168]. b, d, and g mark how listeners identified the consonants. (b) A reinterpretation of [166] with added virtual F2 traces in red. In A, F2 is fully continuous; in B, the initial portion of F2 is silenced.
Brainsci 15 00033 g012
Figure 13. Spectrogram of the Mandarin phrase “比麻黎伪善” /bǐ má lí wěi shàn/ [more hypocritical than Ma Li] [155], with broad phonetic transcriptions and pitch tracts (in blue). In both panels, C, V, and T stand for consonant, vowel, and tone, respectively. In (A), the segmentation is conventional [63,64]. The segmentation of /w/ is based on [65]. In (B) the segmentation is based on the synchronization hypothesis. In (C) the segmentation is based on the diphone principle [202,203].
Figure 13. Spectrogram of the Mandarin phrase “比麻黎伪善” /bǐ má lí wěi shàn/ [more hypocritical than Ma Li] [155], with broad phonetic transcriptions and pitch tracts (in blue). In both panels, C, V, and T stand for consonant, vowel, and tone, respectively. In (A), the segmentation is conventional [63,64]. The segmentation of /w/ is based on [65]. In (B) the segmentation is based on the synchronization hypothesis. In (C) the segmentation is based on the diphone principle [202,203].
Brainsci 15 00033 g013
Table 1. Motor synchrony vs. entrainment.
Table 1. Motor synchrony vs. entrainment.
PropertyMotor SynchronyEntrainment
Synchrony in a single cycle?YesN/A
Speed of achieving synchronyImmediate (1–2 cycles)Many cycles
Similarity in natural frequency?NoYes
In-synch out-synch undulation?NoYes
Under central/shared control?YesNo
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Y. Syllable as a Synchronization Mechanism That Makes Human Speech Possible. Brain Sci. 2025, 15, 33. https://doi.org/10.3390/brainsci15010033

AMA Style

Xu Y. Syllable as a Synchronization Mechanism That Makes Human Speech Possible. Brain Sciences. 2025; 15(1):33. https://doi.org/10.3390/brainsci15010033

Chicago/Turabian Style

Xu, Yi. 2025. "Syllable as a Synchronization Mechanism That Makes Human Speech Possible" Brain Sciences 15, no. 1: 33. https://doi.org/10.3390/brainsci15010033

APA Style

Xu, Y. (2025). Syllable as a Synchronization Mechanism That Makes Human Speech Possible. Brain Sciences, 15(1), 33. https://doi.org/10.3390/brainsci15010033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop