Hostname: page-component-cd9895bd7-jkksz Total loading time: 0 Render date: 2024-12-22T07:07:30.789Z Has data issue: false hasContentIssue false

Underspecification in time

Published online by Cambridge University Press:  15 December 2022

William Idsardi*
Affiliation:
University of Maryland, College Park, USA
Rights & Permissions [Opens in a new window]

Abstract

Substance-free phonology or SFP (Reiss 2017) has renewed interest in the question of abstraction in phonology. Perhaps the most common form of abstraction through the absence of substance is underspecification, where some aspects of speech lack representation in memorized representations, within the phonology or in the phonetic implementation (Archangeli 1988, Keating 1988, Lahiri and Reetz 2010 among many others). The fundamental basis for phonology is argued to be a mental model of speech events in time, following Raimy (2000) and Papillon (2020). Each event can have properties (one-place predicates that are true of the event), which include the usual phonological features, and also structural entities for extended events like moras and syllables. Features can be bound together in an event, yielding segment-like properties. Pairs of events can be ordered in time by the temporal logic precedence relation represented by ‘<’. Events, features and precedence form a directed multigraph structure with edges in the graph interpreted as “maybe next”. Some infant bimodal speech perception results are examined using this framework, arguing for underspecification in time in the developing phonological representations.

Résumé

Résumé

La phonologie sans substance ou SFP (Reiss 2017) a renouvelé l'intérêt pour la question de l'abstraction en phonologie. La forme la plus commune d'abstraction par l'absence de substance est peut-être la sous-spécification, où certains aspects de la parole manquent de représentation dans les représentations mémorisées, au sein de la phonologie ou dans l'implémentation phonétique (Archangeli 1988, Keating 1988, Lahiri and Reetz 2010 parmi beaucoup d'autres). La base fondamentale de la phonologie est un modèle mental des événements de la parole dans le temps, suivant Raimy (2000) et Papillon (2020). Chaque événement peut avoir des propriétés (prédicats monovalents qui sont vrais pour l’événement) qui incluent les traits phonologiques classiques, et aussi des entités structurelles pour les événements étendus comme les mores et les syllabes. Les traits peuvent être liés entre eux dans un événement, ce qui donne des propriétés de type segment. Les paires d’événements peuvent être ordonnées dans le temps par la relation logique temporelle de précédence, soit <. Les événements, les traits et la précédence forment une structure multigraphique dirigée, les arêtes du graphe étant interprétées comme « peut-être le prochain ». Certains résultats de la perception bimodale de la parole chez le nourrisson sont examinés à l'aide de ce cadre, ce qui plaide en faveur d'une sous-spécification temporelle dans les représentations phonologiques en développement.

Type
Article
Copyright
Copyright © Canadian Linguistic Association/Association canadienne de linguistique 2022

1. Introduction

Substance-free phonology or SFP (Reiss Reference Reiss, Hannahs and Bosch2017) has renewed interest in the question of abstraction in phonology. Perhaps the most common form of abstraction through the absence of substance is underspecification, where some aspects of speech lack representation in memorized representations, within the phonology or in the phonetic implementation (Archangeli Reference Archangeli1988, Keating Reference Keating1988, Lahiri and Reetz Reference Lahiri and Reetz2010 among many others). The main locus of underspecification has been features within segments, but in this article I will build toward an argument saying that analyses should also allow for the underspecification of temporal order between items – that not all elements have a definite order established between them at all levels of representation. To do this, I will begin with a proposal for a general framework for phonological representations in terms of directed graphs of phonological events (the nodes or vertices of the graph) where features are monadic properties (one-place predicates) of events and where pairs of events can stand in a dyadic relation (two-place predicate) of temporal precedence (the directed edges of the graph). Modalities other than speech which make use of spatial relationships (sign languages, Braille, etc.) will also need further spatial relations in addition to the temporal precedence relation. It is also possible that additional spatial relations would be useful for speech, but the ability to perceive speech from monaural recordings suggests that spatial localization is not a significant factor in speech perception, whereas the location of a handshape can be contrastive in sign language (Brentari Reference Brentari1998). Importantly, this proposal allows for underspecification of features (that is, not all events have all monadic properties) and of time (that is, not all pairs of events have a dyadic relation of precedence).

I am completely in favour of abstract phonology, that is, phonology not completely determined by the phonetics, in which there are interesting “slippages” between the representations for speech in the memory, action and perception systems. This is the position taken in Avery and Idsardi (Reference Avery, Idsardi and Hall2001), for instance. However, in order to be usable, any model has to have quasi-veridical mappings between the systems, that is, constrained variation in the mapping between auditory cues and phonological categories, for instance. So, one understanding of SFP would be that these relationships are instead unconstrained and learned in all cases. (Most of the discussion of this issue has centred on phonological features, whereas this article will focus instead on time and temporal relationships.) For features, results from neonates and young infants have consistently revealed various biases in the systems relating memory, action and perception, and I will discuss one of these studies (Baier et al. Reference Baier, Idsardi and Lidz2007), below. Idemaru and Holt (Reference Idemaru and Holt2014) also provide extremely relevant evidence for constrained variation, in two ways. First, voice onset time (VOT) serves as the primary cue for English stops, with F0 on the following vowel serving only as a secondary cue. Second, while it is possible, under brief exposures to a novel idiolect, to reduce the reliance on the secondary cue, and even extinguish it – and therefore this cue is, to a certain extent, malleable and plastic –, in Idemaru and Holt's experiments, it was not possible to reverse the relationship between F0 and category membership. That is, higher F0 on the following vowel cannot be made to signal “voicing”, nor lower F0 signal “voicelessness”.

If phonological category formation is as labile and arbitrary as SFP proponents seem to suggest (contra studies like Heffner et al. Reference Heffner, Idsardi and Newman2019), then one potentially fruitful area of research would be the influence of orthography on speech production and perception in literate speakers. There certainly have been claims for (Rastle et al. and Davis Reference Rastle, McCormick, Bayliss and Davis2011) and against (Damian and Bowers Reference Damian and Bowers2009, Mitterer and Reinisch Reference Mitterer and Reinisch2015) orthographic effects in speech production and perception tasks not involving overt use of orthography. (There is robust evidence for orthographic effects in tasks that do overtly involve orthography, e.g., Li et al. Reference Li, Wang and Idsardi2015.)

2. Events, features and precedence

In my view, phonology must interface effectively between at least three components of the mind/brain: memory (long-term memory for the lexicon, short-term memory for what Baddeley (Reference Baddeley1986) has termed the ‘phonological loop’, and episodic memory for particular people and situations), action (for the motor control of the oral, manual and other articulators) and perception (at least auditory, visual, and tactile), which together constitute a memory-action-perception loop (Poeppel and Idsardi Reference Poeppel, Idsardi, Gaskell and Zwitserlood2011). A small portion of this system is shown schematically in Figure 1.

Figure 1: A memory-action-perception loop.

In speech production, items are drawn from long-term memory and assembled into a motor plan to be executed by the motor system (downwards along the left side of Figure 1). In speech perception, external sensations are perceived and ultimately recognized as conglomerations of items in long-term memory (upwards along the right side of Figure 1). The motor system is coupled to sensation with efference copies and feedback (see Gallistel Reference Gallistel1980 for a review), and a similar, more abstract relation between action and perception has been proposed in the analysis-by-synthesis model of speech perception (Halle and Stevens Reference Halle and Stevens1963, Bever and Poeppel Reference Bever and Poeppel2010).

I will argue here for a view that part of the fundamental basis for phonology is a mental model of individuated speech events in time, following Raimy (Reference Raimy2000) and Papillon (Reference Papillon2020) (and also strongly influenced by Pietroski Reference Pietroski2005, Reference Pietroski2018). Each event can have properties (monadic predicates that are true of that event), which we will equate with the usual phonetic features and with structural entities like moras and syllables. The idea is that moras and syllables cover intervals of time in a continuous model, so these intervals are then single events in the discrete model, which are in parallel temporally with the segmental events that are “contained” in the intervals for the moras or syllables. Following Jakobson et al. (Reference Jakobson, Fant and Halle1951), the features provide the “glue” between action and perception (because features must have definitions in terms of both action/articulation and perception/audition), and I also claim, following Poeppel et al. (Reference Poeppel, Idsardi and van Wassenhove2008) (but with much less evidence), that features constitute a basic element of storage in long-term memory. An individual event can have multiple features; equivalently, features can be bound together in an event, yielding segment-like properties (like Firthian sounds; see Firth Reference Firth1948, Kazanina et al. 2017), or events can contain relatively few features, even just a single feature, leading to autosegmental-like behaviours (akin to Firthian prosodies).

The other necessary concept is at least one temporal relation between events: precedence. For sign languages, the events are also located in signing space, and so we will require spatial relations in addition to the temporal relation. Similarly, orthographic systems will require a specification of the relevant spatial relations, which may require more than one spatial dimension, as is the case for Hangul, and a mapping between spatial and temporal relations, as in Braille reading where the fingertip scans an embossed line of text. The kinds of sensible temporal relations depend greatly on the model of time that is employed in the model. One explicit model is a continuous timeline, as employed in Articulatory Phonology (Browman and Goldstein Reference Browman and Goldstein1989) and Time Map Phonology (Carson-Berndsen 1998); see also Bird and Klein (Reference Bird and Klein1990). Events in such models are construed as continuous intervals of time, and admit for temporal relations such as overlap in time (sharing some sub-interval of the timeline), precedence (i.e., notions of before and after), and whether events have a functional form (as in Articulatory Phonology, or in the musical tradition of analysis into attack, sustain and release; see Françoise et al. Reference Françoise, Caramiaux and Bevilacqua2012), and higher-order temporal relations like phasing. These ideas also pervade continuous signal processing models, especially continuous wavelet models which provide an event-based alternative to Fourier analysis (e.g., Suni et al. Reference Suni, Šimko, Aalto and Vainio2017). The continuous-time gestural score for “palm” from Browman and Goldstein (Reference Browman and Goldstein1989: 212) is shown in Figure 2.

Figure 2: Gestural score for “palm”, Browman and Goldstein Reference Browman and Goldstein1989: 212.

It is commonly observed that phonetic parameters in articulation and audition can have a wide range of continuous values, whereas phonological representations typically quantize (or discretize) these values into a small number of options, and also reduce the dimensionality of the problem by picking a primary dimension for the phonological parameter, or by grouping together several phonetic dimensions through the use of phonological latent variables. These techniques of discretization, dimensionality reduction and latent variables are also common techniques in machine learning more generally (Murphy Reference Murphy2012). Browman and Goldstein (Reference Browman and Goldstein1989) also provide a discrete version of their gestural score, in what they term “point notation”. A point notation diagram, again for “palm”, is shown in Figure 3. Browman and Goldstein also endorse, at least implicitly, a mapping between Figure 2 and Figure 3, that is, between continuous and discrete representations.

Figure 3: Point notation diagram for “palm”, Browman and Goldstein Reference Browman and Goldstein1989: 213.

Generative phonology has also taken a much more discrete approach to phonological representations, with features and segments conceived as points in time, albeit with the temporal aspects usually more implicit than explicit, indicated by position in a string, or by spatial position in a diagram. The relevant, but often vague, temporal notions have been something like a containment relation (two features contained in the same segment or two segments contained in an onset), and a synchronization relation in the association lines of Autosegmental Phonology (Goldsmith Reference Goldsmith1976, Sagey Reference Sagey1988, Coleman and Local Reference Coleman and Local1991).

There is not sufficient space in this article to discuss the general issue of continuous and discrete models of time, or physical models of time, which are difficult and contentious issues in physics and metaphysics. Moreover, whatever the physical nature of time, human mental models of time are probably only quasi-veridical, agreeing in some ways with the physical world, but failing to be fully veridical with it (and hence models). The idea that perception in part chunks the perceived world into discretized pieces has a long history, and much recent support in multiple time-scale models for speech (Luo and Poeppel Reference Luo and Poeppel2012, Teng et al. Reference Teng, Tian and Poeppel2016). I find this view compelling, but it does not entail that discrete representations are the only data structures used by the mind/brain for storing, acting and perceiving speech. That is, it is certainly possible, even likely, that there are mappings between continuous and discrete models in terms of discretization (continuous to discrete in perception) and production (as in the motor articulation of an event in terms of prolonged periods for the attack, sustain and release) akin to those in Browman and Goldstein (Reference Browman and Goldstein1989).

Importantly, most of these models assume a linear notion of time (implied by the term timeline). However, following Raimy (Reference Raimy2000) and Papillon (Reference Papillon2020), the view adopted here is that phonological time can have loops, and that the appropriate representation for temporal order is a directed graph (or, more properly, a multigraph, see, e.g., Chartrand and Zhang Reference Chartrand and Zhang2012), which is allowed to have cycles, which are used to analyze repeated event sequences such as reduplication in language, and repetition in music (Bamberger Reference Bamberger1991, Idsardi and Raimy Reference Idsardi, Raimy, Vaux and Nevins2008). There is one large difference between Raimy (Reference Raimy2000) and Papillon (Reference Papillon2020): Raimy (Reference Raimy2000) retains the framework of Autosegmental Phonology and association lines, whereas Papillon (Reference Papillon2020) abandons association lines in favour of a network flow model of simultaneous articulations, compare Figure 4a and Figure 4b.

Figure 4a. [ãŋẽn-ãŋẽn] Raimy Reference Raimy2000: 18.

Figure 4b: [ãŋẽn-ãŋẽn] reinterpreted via Papillon Reference Papillon2020.

These are examples of the overapplication of progressive nasalization in reduplication in forms such as [aŋẽn] => [ãŋẽn-ãŋẽn]. In both cases [+nasal] comes to have “scope” over the entire form. This is handled in Raimy (Reference Raimy2000) by association lines between [+nasal] and the individual segment-sized events; in comparison, Papillon (Reference Papillon2020) handles such cases by placing the [+nasal] event in parallel to the entire stream of segmental events.

I will adopt here Papillon's conjecture that a single temporal relation suffices, and that this is the precedence relation, which is the basic relation in discrete temporal logic (van Benthem Reference van Benthem1983). The association lines in Autosegmental Phonology – interpreted as a synchronization relation between events – is an obvious possible addition to the precedence relation, but Papillon's model employs a strategy of reducing synchronization effects to a lack of precedence specification between events. In the discrete representations, the directed edges can also be interpreted as something akin to “maybe next” (◊N in modal temporal logic, Goranko and Rumberg Reference Goranko and Rumberg2020; there is no simple successor relation in continuous time). Directed edges can occur both within a “tier” or “stream” of events, and across tiers/streams. This has the consequence that the production of a motor plan from a phonological graph is not (as in Raimy Reference Raimy2000) finding a path in a graph from beginning (#) to end (%), but rather finding the flow through a network (Barabási Reference Barabási2016), because parallel edges can be traversed simultaneously (e.g., [+nasal] in Figure 4b).

In summary, the present proposal is that spoken language has an <E, F, P> structure for phonological representations, consisting of a set of events (E, abstract points in time), a set of monadic predicates over events (F, features, e.g., a feature f could be true of event e, f(e)), and a (potentially incomplete) binary relation, <, of temporal precedence between events (such that e1 < e2 is true if e1 happens before e2). I will call this the EFP model.

This sketch has been rather abstract and terse, but I hope that the small case study that is examined in the following sections will make the relevant points clearer.

3. A puzzle in early infant bimodal speech perception

Infants can show a puzzling range of phonetic abilities and deficits in comparison with adults, out-performing adults on many phonetic perception tasks while lagging behind in other ways. Some null results using one procedure can be overturned with more sensitive procedures and some contrasts are “better” than others in terms of effect size and various acoustic or auditory measures of similarity (Sundara et al. Reference Sundara, Ngon, Skoruppa, Feldman, Onario, Morgan and Peperkamp2018). There are additional oddities in the infant speech perception literature, including the fact that the syllabic stimuli generally need to be much longer than the average syllable durations in adult speech (often twice as long). One persistent idea is that infants start with a syllable-oriented perspective and only later move to a more segment-oriented one (Bertoncini and Mehler Reference Bertoncini and Mehler1981, MacNeilage Reference MacNeilage2008), and that in some languages adults still have a primary orientation for syllables, at least for some speech production tasks (O'Seaghdha et al. Reference O'Seaghdha, Chen and Chen2010, though see various replies, e.g., Qu et al. Reference Qu, Damian and Kazanina2012, Li et al. Reference Li, Wang and Idsardi2015).

Baier et al. (Reference Baier, Idsardi and Lidz2007) investigated audio-visual (AV) integration in two-month-old infants, replicating and extending the results of Kuhl and Meltzoff (Reference Kuhl and Meltzoff1982) and Patterson and Werker (Reference Patterson and Werker2003) on bimodal speech perception in infants. Infants were presented with a single audio track along with two synchronized silent videos of the same person (Rebecca Baier) articulating single syllables, presented on a large TV screen, and the infants’ looks to each face were tabulated and analyzed. The infants were able to visually distinguish and match articulating faces with synchronized audio for [a], [i] and [u], replicating the previous findings. Taking this one step further, the infants could also detect dynamic syllables, matching faces with audio when tested on [wi] vs. [i]. However, we were subsequently puzzled by infants’ inability to distinguish between [wi] and [ju] which, moreover, correspond to the very high frequency English words “we” and “you”. The overall proportion of matching looks was 0.52 for [wi] (n = 20, p > 0.5, n.s.) and 0.49 for [ju] (n = 18, p > 0.5, n.s.); neither was significantly different from chance, and they were not different from each other (unpaired n = 21, p > 0.69 n.s., paired n = 17, p > 0.71 n.s., including only those infants who completed both conditions). Furthermore, when they were presented with [ju] audio alongside [wi] and [i] faces, they matched the [ju] audio with the [wi] video. These behaviours are at least consistent with a syllable-oriented point of view in terms of features: infants hear a dynamic syllable with something [round] and something [front] in it, but they cannot tell the relative order of the [front] and [round] events. This is also consistent with the relatively poor abilities of infants to detect differences in serial order (Lewkowicz Reference Lewkowicz2004, Warren Reference Warren2008). Importantly, this is not to say that infants cannot hear a difference between [wi] and [ju], which is strongly signaled acoustically by the trajectory of the second formant: rising F2 in [wi] and falling F2 in [ju]. I am confident that dishabituation experiments on infants would succeed on this contrast, which is a relatively easy one auditorily. The anticipated contrast in behaviour reinforces the importance of the choice of experimental task in evaluating infant abilities which can span a number of perceptual levels of representation. For various mundane reasons we did not pursue additional bimodal infant speech perception experiments at the time.

I believe that Papillon's model now provides a formal way of understanding this curious finding of /wi~ju/ because the model allows for representations which are underspecified in time, a possibility that is difficult or even impossible to achieve with other phonological theories. That is, this is a representational advance akin to those achieved in Goldsmith (Reference Goldsmith1976) and Raimy (Reference Raimy2000) which broadened and refined our notions of temporal relations in phonology.

4. A brief analysis

I propose that the EFP model provides a succinct formal account for these puzzling results, finding an equivalence for the infants in /wi/ ~ /ju/, and successfully encodes the informal understanding outlined above. In the EFP framework, we can capture an ambiguous /wi~ju/ syllable schematically (suppressing irrelevant details) as in Figure 5.

Figure 5: EFP representation of an ambiguous /wi~ju/ syllable.

The relative ordering of [front] and [round] is underspecified at this point, as is the temporal extent of the events, but their temporal relationship is not altogether unknown, for they occur in the same syllable. That is, [front] and [round] are not just floating as in the analysis of floating tones in Autosegmental Phonology; their temporal representation is only partially under-determined, perhaps more akin to Firthian prosodies covering an entire syllable. The ultimate ability to discriminate between /wi/ and /ju/ amounts to learning and incorporating the relative ordering of [front] and [round]. When [round] < [front], that is the developing representation for /wi/; when [front] < [round], that is the developing representation for /ju/. Acquiring this kind of serial order knowledge between different features might be fairly difficult, as it is likely that [front] and [round] are initially segregated into different auditory streams (Bregman Reference Bregman1990), and order perception across streams is worse than that within streams. Moreover, it is a contingent fact of English phonology that the vowel system does not allow a [front, round] combination, so we expect some differences in the developmental trajectory of infants learning French or German where the [front, round] combination is licit. One possibility is that the learner would be driven to look for additional temporal relations when the temporally underspecified representations incorrectly predict such cases of “homophony”, that is, when they are presented with evidence that /wi/ ≠ /ju/ and therefore conclude that distinct phonological representations are required. If we pursue this idea generally, the EFP graphs will gradually become somewhat more segment-oriented as additional precedence relations are resolved, economizing on the number of events, and leading to later representations, as in Figure 6 for /ju/, where some events have been fused into feature bundles such that [front] < [round] holds.

Figure 6: EFP representation for /ju/ after feature bundling.

So, the general proposal made here is that the developing knowledge of the relative order of features is initially poor and underspecified for order between features in different “streams” (Bregman Reference Bregman1990). A more general conclusion is that features are not learned from earlier representations involving complete phonetic segments (Mielke Reference Mielke2008, Odden 2019) but rather that features are gradually combined during development to form more nearly segment-sized units. That is, rather than starting with whole segments as the basic entities and then splitting them up into features, instead events with individual features loosely coupled with “longer” events such as syllables (Poeppel Reference Poeppel2003) are the starting point, and segments can be constructed by combining events together, as in the fusion of C with [front] in Figure 6. Ultimately this is an empirical question, but the view developed here suggests that temporal precision in speech perception should improve during language development.

References

Archangeli, Diana. 1988. Aspects of underspecification theory. Phonology 5(2): 183207.CrossRefGoogle Scholar
Avery, Peter, and Idsardi, William J.. 2001. Laryngeal dimensions, completion and enhancement. In Distinctive feature theory, ed. Hall, T. Alan, 4170. Berlin: Mouton de Gruyter.Google Scholar
Baddeley, Alan. 1986. Working memory. Oxford: Oxford University Press.Google ScholarPubMed
Baier, Rebecca, Idsardi, William J., and Lidz, Jeffrey. 2007. Two-month-olds are sensitive to lip rounding in dynamic and static speech events. In Proceedings of the International Conference on Auditory-Visual Speech Processing, L6–2. <https://www.isca-speech.org/archive/avsp_2007/baier07_avsp.html>>Google Scholar
Bamberger, Jeanne. 1991. The mind behind the musical ear. Cambridge, MA: Harvard University Press.Google Scholar
Barabási, Albert-László. 2016. Network science. Cambridge: Cambridge University Press.Google Scholar
van Benthem, Johan F. K. 1983. The logic of time: A model-theoretic investigation into the varieties of temporal ontology and temporal discourse. Dordrecht: D. Reidel.CrossRefGoogle Scholar
Bertoncini, Josiane, and Mehler, Jacques. 1981. Syllables as units in infant speech perception. Infant Behavior and Development 4: 247260.CrossRefGoogle Scholar
Bever, Thomas. G., and Poeppel, David. 2010. Analysis by synthesis: A (re-) emerging program of research for language and vision. Biolinguistics 4(2–3): 174200.CrossRefGoogle Scholar
Bird, Steven, and Klein, Ewan. 1990. Phonological events. Journal of Linguistics 26(1): 3356.CrossRefGoogle Scholar
Bregman, Albert S. 1990. Auditory scene analysis. Cambridge, MA: MIT Press.CrossRefGoogle Scholar
Brentari, Diane. 1998. A prosodic model of sign language phonology. Cambridge, MA: MIT Press.Google Scholar
Browman, Catherine P., and Goldstein, Louis. 1989. Articulatory gestures as phonological units. Phonology 6(2): 201251.CrossRefGoogle Scholar
Carson-Berndsen, Julie. 1988. Time map phonology: Finite state models and event logics in speech recognition. Berlin: Springer.Google Scholar
Chartrand, Gary, and Zhang, Ping. 2012. A first course in graph theory. New York: Dover Publications.Google Scholar
Coleman, John, and Local, John. 1991. The “No Crossing Constraint” in Autosegmental Phonology. Linguistics and Philosophy 14(3): 295338.CrossRefGoogle Scholar
Damian, Markus F., and Bowers, Jeffrey S.. 2009. Assessing the role of orthography in speech perception and production: Evidence from picture–word interference tasks. European Journal of Cognitive Psychology 21(4): 581598.CrossRefGoogle Scholar
Firth, John R. 1948. Sounds and prosodies. Transactions of the Philological Society. Philological Society 47(1): 127152.CrossRefGoogle Scholar
Françoise, Jules, Caramiaux, Baptiste, and Bevilacqua, Frédéric. 2012. A hierarchical approach for the design of gesture-to-sound mappings. In Proceedings of the 9th Sound and Music Computing Conference, ed. Stefania Serafin, 233240.Google Scholar
Gallistel, Randy. 1980. The organization of action: A new synthesis. Hillsdale: Lawrence Erlbaum Associates.Google Scholar
Goranko, Valentine, and Rumberg, Antje. 2020. Temporal logic. The Stanford Encyclopedia of Philosophy. <https://plato.stanford.edu/archives/sum2020/entries/logic-temporal/>>Google Scholar
Goldsmith, John. 1976. Autosegmental phonology. Doctoral dissertation, Massachusetts Institute of Technology.Google Scholar
Halle, Morris, and Stevens, Kenneth N.. 1963. Speech recognition: A model and a program for research. IRE Transactions on Information Theory 8: 155159.CrossRefGoogle Scholar
Heffner, Christopher C., Idsardi, William J., and Newman, Rochelle S.. 2019. Constraints on learning disjunctive, unidimensional auditory and phonetic categories. Attention, Perception, and Psychophysics 81(4): 958980.CrossRefGoogle ScholarPubMed
Idemaru, Kaori, and Holt, Lori L. 2014. Specificity of dimension-based statistical learning in word recognition. Journal of Experimental Psychology: Human Perception and Performance 40(3): 10091021.Google ScholarPubMed
Idsardi, William J., and Raimy, Eric. 2008. Reduplicative economy. In Rules and constraints in contemporary phonological theory, ed. Vaux, Bert and Nevins, Andrew, 149184. Oxford: Oxford University Press.Google Scholar
Jakobson, Roman, Fant, Gunnar, and Halle, Morris. 1951. Preliminaries to speech analysis. Cambridge, MA: MIT Press.Google Scholar
Kazanina, Nina, Bowers, Jeffrey S., and Idsardi, William J.. 2018. Phonemes: Lexical access and beyond. Psychonomic Bulletin and Review 25(2): 560585.CrossRefGoogle ScholarPubMed
Keating, Patricia. 1988. Underspecification in phonetics. Phonology 5(2): 275292.CrossRefGoogle Scholar
Kuhl, Patricia K., and Meltzoff, Andrew N.. 1982. The bimodal perception of speech in infancy. Science 218: 11381141.CrossRefGoogle ScholarPubMed
Lahiri, Aditi, and Reetz, Henning. 2010. Distinctive features: Phonological underspecification in representation and processing. Journal of Phonetics 38(1): 4459.CrossRefGoogle Scholar
Lewkowicz, David J. 2004. Perception of serial order in infants. Developmental Science 7(2): 175184.CrossRefGoogle ScholarPubMed
Li, Chuchu, Wang, Min, and Idsardi, William J.. 2015. The effect of orthographic form-cuing on the phonological preparation unit in spoken word production. Memory and Cognition 43(4): 563578.CrossRefGoogle ScholarPubMed
Luo, Huan, and Poeppel, David. 2012. Cortical oscillations in auditory perception and speech: Evidence for two temporal windows in human auditory cortex. Frontiers in Psychology 3: 170.CrossRefGoogle ScholarPubMed
MacNeilage, Peter F. 2008. The origin of speech. Oxford: Oxford University Press.Google Scholar
Mielke, Jeff. 2008. The emergence of distinctive features. Oxford: Oxford University Press.Google Scholar
Mitterer, Holger, and Reinisch, Eva. 2015. Letters don't matter: No effect of orthography on the perception of conversational speech. Journal of Memory and Language 85: 116134.CrossRefGoogle Scholar
Murphy, Kevin P. 2012. Machine learning: A probabilistic perspective. Cambridge, MA: MIT Press.Google Scholar
Odden, David. 2022. Radical substance free phonology and feature learning. Journal of Canadian Linguistics 67(4): 500551.CrossRefGoogle Scholar
O'Seaghdha, Padraig G., Chen, Jenn-Yeu, and Chen, Train-Min. 2010. Proximate units in word production: Phonological encoding begins with syllables in Mandarin Chinese but with segments in English. Cognition 115(2): 282302.CrossRefGoogle ScholarPubMed
Papillon, Maxime. 2020. Precedence and the lack thereof: Precedence-relation-oriented phonology. Doctoral dissertation, University of Maryland.Google Scholar
Patterson, Michelle L., and Werker, Janet F.. 2003. Two-month-old infants match phonetic information in lips and voice. Developmental Science 6(2): 193198.CrossRefGoogle Scholar
Pietroski, Paul. 2005. Events and semantic architecture. Oxford: Oxford University Press.Google Scholar
Pietroski, Paul. 2018. Conjoining meanings: Semantics without truth values. Oxford: Oxford University Press.CrossRefGoogle Scholar
Poeppel, David. 2003. The analysis of speech in different temporal integration windows: Cerebral lateralization as ‘asymmetric sampling in time’. Speech Communication 41(1): 245255.CrossRefGoogle Scholar
Poeppel, David, Idsardi, William J., and van Wassenhove, Virginie. 2008. Speech perception at the interface of neurobiology and linguistics. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences 363(1493): 10711086.CrossRefGoogle ScholarPubMed
Poeppel, David, Idsardi, William J.. 2011. Recognizing words from speech: The perception-action-memory loop. In Lexical representation: A multidisciplinary approach, ed. Gaskell, Gareth and Zwitserlood, Pienie, 171196. Berlin: Mouton de Gruyter.CrossRefGoogle Scholar
Qu, Qingqing, Damian, Markus F., and Kazanina, Nina. 2012. Sound-sized segments are significant for Mandarin speakers. Proceedings of the National Academy of Sciences 109(35): 1426514270.CrossRefGoogle ScholarPubMed
Raimy, Eric. 2000. The phonology and morphology of reduplication. Berlin: Mouton de Gruyter.CrossRefGoogle Scholar
Rastle, Kathleen, McCormick, Samantha F., Bayliss, Linda, and Davis, Colin J.. 2011. Orthography influences the perception and production of speech. Journal of Experimental Psychology: Learning, Memory, and Cognition 37(6): 15881594.Google ScholarPubMed
Reiss, Charles. 2017. Substance free phonology. In The Routledge handbook of phonological theory, ed. Hannahs, S. J. and Bosch, Anna R. K., 425452. London: Routledge.CrossRefGoogle Scholar
Sagey, Elizabeth. 1988. On the ill-formedness of crossing association lines. Linguistic Inquiry 19(1): 109118.Google Scholar
Sundara, Megha, Ngon, Céline, Skoruppa, Katrin, Feldman, Naomi H., Onario, Glenda Molina, Morgan, James L., and Peperkamp, Sharon. 2018. Young infants’ discrimination of subtle phonetic contrasts. Cognition 178: 5766.CrossRefGoogle ScholarPubMed
Suni, Antti, Šimko, Juraj, Aalto, Daniel, and Vainio, Martti. 2017. Hierarchical representation and estimation of prosody using continuous wavelet transform. Computer Speech and Language 45: 123136.CrossRefGoogle Scholar
Teng, Xianbing, Tian, Xing, and Poeppel, David. 2016. Testing multi-scale processing in the auditory system. Scientific Reports 6: 34390.CrossRefGoogle ScholarPubMed
Warren, Richard M. 2008. Auditory perception: An analysis and synthesis. 3rd ed. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
Figure 0

Figure 1: A memory-action-perception loop.

Figure 1

Figure 2: Gestural score for “palm”, Browman and Goldstein 1989: 212.

Figure 2

Figure 3: Point notation diagram for “palm”, Browman and Goldstein 1989: 213.

Figure 3

Figure 4a. [ãŋẽn-ãŋẽn] Raimy 2000: 18.

Figure 4

Figure 4b: [ãŋẽn-ãŋẽn] reinterpreted via Papillon 2020.

Figure 5

Figure 5: EFP representation of an ambiguous /wi~ju/ syllable.

Figure 6

Figure 6: EFP representation for /ju/ after feature bundling.