Defectivity in Kinande

[This is part of a series of defectivity case studies.]

I have already written a bit about reduplication in Kinande; it too is an example of inflectional defectivity, and here I’ll focus on that fact.

In this language, most verbs participate in a form of reduplication with the semantics of roughly ‘to hurriedly V’ or ‘to repetitively V’. Mutaka & Hyman (1990; henceforth MH), argue that the reduplicant is a bisyllabic prefix. For instance, the reduplicated form of e-ri-gend-a ‘to leave’ is e-ri-gend-a-gend-a ‘to leave hurriedly’, with the reduplicant underlined. (In MH’s terms, e- is the “augment”, -ri the “prefix”, and -a is the “final vowel” morpheme.)

Certain verbal suffixes, known to Bantuists as extensions, may also be found in the reduplicant when the reduplicant would otherwise be less than bisyllabic. For instance, the passive suffix, underlyingly /-u-/, surfaces as [w] and is copied by reduplication. Thus for the verb root hum ‘beat’ the passive e-ri-hum-w-a reduplicates as e-ri-hum-w-a-hum-w-a. More interesting is there are “unproductive” (MH’s term) extensions.1 Verbs bearing these extensions rarely have a compositional semantic relationship with their unextended form (if an unextended verb stem exists at all). For instance, whereas luh-uk-a ‘take a rest’ may be semantically related to luh-a ‘be tired’, but there is no unextended *bát-a to go with bát-uk-a ‘move’.

Interesting things happen when we try to reduplicate unproductivity extended monosyllabic verb roots. For some such verbs, the extension is not reduplicated; e.g., e-rí-bang-uk-a ‘to jump about’ has a reduplicated form e-rí-bang-a-bang-uk-a. This is the same behavior found for “productive” extensions. For others, the extension is reduplicated, producing a trisyllabic—instead of the normal bisyllabic—reduplicant; e.g., e-ri-hurut-a ‘to snore’ has a reduplicated form e-ri-hur-ut-a-hur-ut-a. Finally, there are some stems—all monosyllabic verb roots with unproductive extensions—which do not undergo reduplication; e.g., e-rí-bug-ul-a ‘to find’ does not reduplicate and neither *e-rí-bug-a-bug-ul-a or *e-rí-bug-ul-a-bug-ul-a exist.

While one could imagine there are certain semantic restrictions on reduplication, like in Chaha, MH make no mention of such restrictions in Kinande. If possible, we should rule out this as a possible explanation for the aforementioned defectivity.

Endnotes

  1. I will segment these with hyphens though it may make sense to regard some unproductive extensions as part of morphologically simplex stems.

References

Mutaka, N. and Hyman, L. M. 1990. Syllables and morpheme integrity in Kinande reduplication. Phonology 7: 73-119.

Defectivity in Polish

[This is part of a series of defectivity case studies.]

Gorman & Yang (2019), following up on a tip from Margaret Borowczyk (p.c.) discuss inflectional gaps in Polish declension. In this language, masculine genitive singular (gen.sg.) are marked either with -a or -u. The two gen.sg. suffixes have a similar type frequency, and neither appears to be more default-like than the other. For instance, both allomorphs are used with loanwords. Because of this, it is generally agreed that the gen.sg. allomorphy is purely arbitrary and must be learned by rote, a process that continues into adulthood (e.g., Dąbrowska 2001, 2005).

Kottum (1981: 182) reports his informants have no gen.sg. for masculine-gender toponyms like Dublin ‘id.’ (e.g., *Dublina/*Dublinu), Göteborg ‘Gothenburg’ and Tarnobrzeg ‘id.’, and Gorman & Yang (2019: 184) report their informants do not have a gen.sg. for words like drut ‘wire’ (e.g., *druta/*drutu, though the latter is prescribed), rower ‘bicycle’, balon ‘baloon’, karabin ‘rifle’, autobus ‘bus’, and lotos ‘lotus flower’.

References

Dąbrowska, E. 2001. Learning a morphological system without a default: The Polish genitive. Journal of Child Language 28: 545-574.
Dąbrowska, E. 2005. Productivity and beyond: mastering the Polish genitive inflection. Journal of Child Language 32:191-205.
Gorman,. K. and Yang, C. 2019. When nobody wins. In F. Rainer, F. Gardani, H. C. Luschützky and W. U. Dressler (ed.), Competition in Inflection and Word Formation, pages 169-193. Springer.
Kottum, S. S. 1981. The genitive singular form of masculine nouns in Polish. Scando-Slavica 27: 179-186.

Defectivity in Chaha

[This is part of a series of defectivity case studies.]

Rose (2000) describes a circumscribed form of defectivity in Chaha, a Semitic language spoken in Ethiopia. Throughout Ethio-Semitic, many verbs have a frequentative formed using a quadriliteral verbal template. Since few verb roots are quadriconsonantal—most are triconsonantal, some are biconsonantal—a sort of reduplication and/or spreading is used to fill in the template. In Tigryina, for instance (p. 318), the frequentative template is of the form CɘCaCɘC. Then, frequentative of the triconsonantal verb root √/grf/ ‘collect’ is [gɘrarɘf], with the root /r/ repeated, and for a biconsonantal verb root like √/ħt/ ‘ask’, the frequentative is [ħatatɘt], with three root /t/s.

Rose contrasts this state of affairs with Chaha. In this language, the frequentative template CɨCɘCɘC cannot be satisfied by a biconsonantal root like √/tʼm/ ‘bend’ or √/Rd/ ‘burn’, and all such verbs lack a frequentative.1 The expected *[tʼɨmɘmɘm] and *[nɨdɘdɘd] are ill-formed, as are all other alternatives. Furthermore, no frequentatives of any sort can be formed with quadriconsonantal roots.

Rose notes that there are often semantic reasons for a verb to lack a frequentative (e.g., stative and resultative verbs are generally not compatible with it), this does not seem applicable here.

Endnotes

  1. As Rose explains: “R represents a coronal sonorant which may be realized as [n] or [r] depending on context…” (p. 317).

References

Rose, S. 2000. Multiple correspondence in reduplication. In Proceedings of the 23rd Annual Meeting of the Berkeley Linguistic Society, pages 315-326.

Defectivity in English

[This is part of a small but growing series of defectivity case studies.]

English lexical verbs can have up to 5 distinct forms, and I am aware of just a few English verbs which are defective. (The following are all my personal judgments.)

  1. I can use begone as an imperative, though it has the form of a past participle (cf. gone and forgone). Is BEGO even a verb lexeme anymore?
  2. Fodor (1972), following Lakoff (1970 [1965]), notes that BEWARE has a limited distribution and never bears explicit inflection. For me, it can occur only as a positive imperative (e.g., beware the dog!), with or without emphatic do. I agree with Fodor that it is also bad under negation, but perhaps for unrelated reasons: e.g., *don’t beware… 
  3. FORGO lacks a simple past: forgo, forgoes, and forgoing are fine, as is the past participle forgone, but *forwent is bad as the preterite/simple past, and *forgoed is perhaps a bit worse.
  4. METHINK can only be used in the 3sg. present active indicative form methinks, and doesn’t allow for an explicit subject.
  5. STRIDE lacks a past participle (e.g., Hill 1976:668, Pinker 1999:136f., Pullum and Wilson 1977:770): *stridden is bad.  The simple past strode cannot be reused here, and I cannot use the regular *strided (under the relevant sense).

References

Fodor, J. D. 1972. Beware. Linguistic Inquiry 3: 528-534.
Hill, A. A. 1976. [Obituary:] Albert Henry Marckwardt. Language 52: 667-681.
Lakoff, G. 1970. Irregularity in Syntax. Holt, Rinehart and Winston.
Pinker, S. 1999. Words and Rules: The Ingredients of Language. Basic Books.
Pullum, G. K. and Wilson, D. 1977. Autonomous syntax and the analysis of auxiliaries. Language 53:741-788.

Deriving the major rule/minor rule distinction

The ability to target underspecified lexemes’ specifications for a rule feature, in which feature-filling is implemented by unification (e.g., Bale et al. 2014), ought to enable us to derive the traditional distinction (e.g., Lakoff 1970) between major rules (those for which non-application is exceptional) and minor rules (those for which application is exceptional), making this distinction purely descriptive of later feature-filling rules inserting unmarked rule features upon lexical insertion.

Let us suppose we have a rule R. Let us suppose that every formative is unified with  {+R} upon lexical insertion. Then, unification will fail only with formatives specified [−R], and these formatives will exhibit exceptional non-application. This describes the parade example of exceptions to a major rule: the failure of trisyllabic shortening in obesity (assuming obese is [−trisyllabic shortening]; see Chomsky & Halle 1968: §4.2.2).

Let us suppose instead that every formative is unified with {−R} upon lexical insertion. Then, unification will fail only with those formatives specified [+R], and these formatives will exhibit exceptional application, assuming they otherwise satisfy the phonological description of rule R. This describes minor rules.

This (admittedly quite sketchy at present) idea seems to address Zonneveld’s (1978: 160f.) concern that Lakoff and contemporaries did not posit any way to encode whether or not a rule was major or minor, except “transderivationally” via inspection of successful derivations. This also places the major/minor distinction—correctly, I think—in the scope of theory of productivity. More on this later.

References

Bale, A., Papillon, M., and Reiss, C. 2014. Targeting underspecified segments: a formal analysis of feature-changing and feature-filling rules. Lingua 148: 240-253.
Chomsky, N. and Halle, M. 1968. Sound Pattern of English. Harper & Row.
Lakoff, G. 1970. Irregularity in Syntax. Holt, Rinehart and Winston.
Zonneveld, W. 1978. A Formal Theory of Exceptions in Generative Phonology. Peter de Ridder.

Linguistics’ contribution to speech & language processing

How does linguistics contribute to speech & language processing? While there exist some “linguist eliminationists”, who wish to process speech audio or text “from scratch” without intermediate linguistic representations, it is generally recognized that linguistic representations are the end goal of many processing “tasks”. Of course some tasks involve poorly-defined, or ill-posed, end-state representations—the detection of hate speech and named entities, neither of which are particularly well-defined, linguistically or otherwise, come to mind—but are driven by apparent business value to be extracted rather than serious goals to understand speech or text.

The standard example for this kind of argument is syntax. It might be the case that syntactic representations are not as useful for textual understanding as was anticipated, and useful features for downstream machine learning can apparently be induced using far simpler approaches, like the masked language modeling task used for pre-training in many neural models. But it’s not as if a terrorist cell of rogue linguists locked NLP researchers in their office until they developed the field of natural language parsing. NLP researchers decided, of their own volition, to spend the last thirty years building models which could recover natural language syntax, and ultimately got pretty good at it, probably getting up to the point where, I suspect, unresolved ambiguities mostly hinge on world knowledge that is rarely if ever made explicit.

Let us consider another example, less widely discussed: the phoneme. The phoneme was discovered in the late 19th century by Baudouin de Courtenay and Kruszewski. It has been around a very long time. In the century and a half since it emerged from the Polish academy, Poland itself has been a congress, a kingdom, a military dictatorship, and a republic (three times), and annexed by the Russian empire, the German Reich, and the Soviet Union. The phoneme is probably here to stay. The phoneme is, by any reasonable account, one of the most successful scientific abstractions in the history of science.

It is no surprise then, that the phoneme plays a major role in speech technologies. Not only did the first speech recognizers and synthesizers make explicit use of phonemic representations (as well as notions like allophones), so did the next five decades worth of recognizers and synthesizers. Conventional recognizers and synthesizers require large pronunciation lexicons mapping between orthographic and phonemic form, and as they get closer to speech, convert these “context-independent” representations of phonemic sequences onto “context-dependent” representations which can account for allophony and local coarticulation, exactly as any linguist would expect. It is only in the last few years that it has even become possible to build a reasonably effective recognizer or synthesizer which doesn’t have an explicit phonemic level of representation. Such models instead use clever tricks and enormous amounts of data to induce implicit phonemic representations instead. We have every reason to suspect these implicit representations are quite similar to the explicit ones linguists would posit. For one, these implicit representations are keyed to orthographic characters, and as I wrote a month ago, “the linguistic analysis underlying a writing system may be quite naïve but may also encode sophisticated phonemic and/or morphemic insights.” If anything, that’s too weak: in most writing systems I’m aware of, the writing system is either a precise phonemic analysis (possibly omitting a few details of low functional load, or using digraphs to get around limitations of the alphabet of choice) or a precise morphophonemic analysis (ditto). For Sapir (1925, et. seq.) this was key evidence for the existence of phonemes! So whether or not implicit “phonemes” are better than explicit ones, speech technologists have converged on the same rational, mentalistic notions discovered by Polish linguists a century and a half ago.

So it is surprising to me that even those schooled in the art of speech processing view the contribution of linguistics to the field in a somewhat negative light. For instance, Paul Taylor, the founder of the TTS firm Phonetic Arts, published a Cambridge University Press textbook on TTS methods in 2009, and while it’s by now quite out of date, there’s no more-recent work of comparable breadth. Taylor spends the first five hundred (!) pages or so talking about linguistic phenomena like phonemes, allophones, prosodic phrases, and pitch accents—at the time, the state of the art in synthesis made use of explicit phonological representations—so it is genuinely a shock to me that Taylor chose to close the book with a chapter (Taylor 2009: ch. 18) about the irrelevance of linguistics. Here are a few choice quotes, with my commentary.

It is widely acknowledged that researchers in the field of speech technology and linguistics do not in general work together. (p. 533)

It may be “acknowledged”, but I don’t think it has ever been true. The number of linguists and linguistically-trained engineers working on FAANG speech products every day is huge. (Modern corporate “AI” is to a great degree just other people, mostly contractors in the Global South.) Taylor continues:

The first stated reason for this gap is the “aeroplanes don’t flap their wings” argument. The implication of this statement is that, even if we had a complete knowledge of how human language worked, it would not help us greatly because we are trying to develop these processes in machines, which have a fundamentally different architecture. (p. 533)

I do not expect that linguistics will provide deep insights about how to build TTS systems, but it clearly identified the relevant representational units for building such systems many decades ahead of time, just as mechanics provided the basis for mechanical engineering. This was true of Kempelen’s speaking machine (which predates phonemic theory, and so had to discover something like it) and Dudley’s voder as well as speech synthesizers in the digital age. So I guess I kind of think that speech synthesizers do flap their wings: parametric, unit selection, hybrid, and neural synthesizers are all big fat phoneme-realization machines. As is standard practice in physical sciences, the simple elementary particles of phonological theory—phonemes, and perhaps features—were discovered quite early on, but it the study of their onotology has taken up the intervening decades. And unlike the physical sciences, us cognitive scientists some day must also understand their epistemology (what Chomsky calls “Plato’s problem”) and ultimately, their evolutionary history (“Darwin’s problem”) too. Taylor, as an engineer, need not worry himself about these further studies, but I think he is being widely uncharitable about the nature of what he’s studying, or the business value of having a well-defined hypothesis space of representations for his team to engineer around in.

Taylor’s argument wouldn’t be complete without a caricature of the generative enterprise:

The most-famous camp of all is the Chomskian [sic] camp, started of course by Noam Chomsky, which advocates a very particular approach. Here data are not used in any explicit sense, quantitative experiments are not performed and little stress is put on explicit description of the theories advocated. (p. 534)

This is nonsense. Linguistic examples are data, in some cases better data than results from corpora or behavioral studies, as the work of Sprouse and colleagues has shown. No era of generativism was actively hostile to behavioral results; as early as the ’60s, generativist-aligned psycholinguists were experimentally testing the derivational theory of complexity and studying morphological decomposition in the lab. And I simply have never found that generativist theorizing lacks for formal explicitness; in phonology, for instance, the major alternative to generativist thinking is exemplar theory—which isn’t even explicit enough to be wrong—and a sort of neo-connectionism—which ought not to work at all given extensive proof-theoretic studies of formal learnability and the formal properties of stochastic gradient descent and backpropagation. Taylor continues to suggest that the “curse of dimensionality” and issues of generalizability prevent application of linguistic theory. Once again, though, the things we’re trying to represent are linguistic notions: machine learning using “features” or “phonemes”, explicit or implicit, is still linguistics.

Taylor concludes with some future predictions about how he hopes TTS research will evolve. His first is that textual analysis techniques from NLP will become increasingly important. Here the future has been kind to him: they are, but as the work of Sproat and colleagues has shown, we remain quite dependent on linguistic expertise—of a rather different and less abstract sort than the notion of the phoneme—to develop these systems.

References

Sapir, E. 1925. Sound patterns in language. Language 1:37-51.
Taylor, P. 2009. Text-to-Speech Synthesis. Cambridge University Press.

Defectivity in Turkish; part 1: monosyllables

[This is part of a small but growing series of defectivity case studies.]

While there are some languages—like Greek or Russian—in which there are dozens or even hundreds of defective lexemes, in most cases defectivity is markedly constrained, conditioned by both morphological class or status and lexical identity. This is somewhat in conflict with models which view defectivity as essentially “absolute phonotactic ungrammaticality” (e.g., Orgun & Sprouse 1999; henceforth OS), since the generalizations about which items are or are not defective are not primarily phonotactic. A good demonstration of the morphological-lexical nature of defectivity comes from Turkish.

As first reported (to my knowledge) by Itô & Hankamer (1989; henceforth IH) Turkish has just a small number of monosyllabic stems. In verbs, one forms the “simple” (active) imperative using the bare stem: e.g., ye ‘eat!’. However, one cannot form a passive imperative of monosyllabic verbs. For instance, for EAT, we would expect *yen (with -n being the expected allomorph of the passive imperative), but this is apparently ill-formed under the appropriate interpretation, with no obvious alternative.1 I say it this way because because yen exists as is the simple imperative ‘conquer!’. As IH note, this shows there is nothing phonotactically wrong with the ill-formed passive imperatives. Another example they give is kon ‘alight! (like a bird)’. Apparently, we would expect it to have a passive imperative homophonous with the simple imperative, but it is ill-formed under this interpretation. However, I find these two examples less than convincing since one could imagine that the homophony with another type of imperative might be implicated in these judgments.

Something similar characterizes certain monosyllabic noun stems. Turkish has apparently borrowed the seven solfège syllables doremi, etc. Of these, six are CV monosyllables, which we would expect to select the /-m/ allomorph of the 1sg. poss. suffix. However, these 1sg. poss. forms are apparently ill-formed; e.g., *do-m ‘my do‘, re-m ‘my re‘, *mi-m ‘my mi‘, and so on. However, one can use these with the other declensional suffixes which produce polysyllabic outputs; e.g., 1pl. poss. domuz ‘our do’. The same facts are true for the names of the letters of the (Latin, post-1928) alphabet: e.g., de ‘the letter d’, but *de-m ‘my letter d’, and so on. OS report however that the one CVC solfège syllable, sol, has a well-formed 1sg. poss.; this selects the /-Im/ allomorph (where /I/ is a high vocalic archiphoneme subject to stem-controlled vowel harmony), which gives us the licit 1sg. poss. solüm [solʲym] ‘my sol‘.2 The same facts hold of the 2sg. poss. ‘your __’, which for CV monosyllables would be realized as /-n/; e.g., *do-n ‘your do‘.

From the above facts IH and OS conclude there is an exceptionless constraint in Turkish such that monosyllabic derived forms produced by the grammar are ill-formed, with no possible “repair”. However, Selin Alkan (p.c.) draws my attention to at least one CV nominal stem which is well-formed in the 1sg. and 2sg,. poss.: su ‘water. For this stem, a [j] glide is inserted between the vowel and the stem, and the stem selects for the -VC allomorphs of the possessive; e.g., su-y-um ‘my water’, su-y-un ‘their water’. This is surprisingly insofar as OS take pains (p. 195f.) to specifically rule out repair by epenthesis in 1sg. poss. forms!

It would be nice to conclude that the only affected lexemes are transparent borrowings, but this does not seem to accord with the evidence from monosyllabic verbs. But the evidence from native stems is really quite weak, and the generalizations are clearly morphological (i.e., the restriction of the constraint to derived environments) and lexical (i.e., the fact that su has an “escape hatch”), something that has largely been ignored in previous attempts to describe defectivity in Turkish.

To move forward on this topic, it would be nice to know the following. How many, if any,  verbs behave like ye or kon, and are any unexpectedly well-formed in the passive imperative? Are there any other forms in the verbal paradigm that show “monosyllabism” gaps? Similarly, how many (if any) defective nouns are there beyond those already mentioned, and how many behave like su?

[h/t: Selin Alkan]

Endnotes

  1. IH note (p. 61, fn. 1) that the passive imperative “are somewhat odd in normal circumstances”. Therefore, they asked their informants to imagine they were directors giving instructions to actors, which apparently helped to render these examples more felicitious.
  2. It seems plausible that -m and -n here are purely-phonological allomorphs of /-Im, -In/ respectively, but I am not sure.

References

Itô, J. and Hankamer, J. 1989. Notes on monosyllabism in Turkish. In J. Itô and J. Runner (ed.), Phonology at Santa Cruz, pages 61-69. Linguistics Research Center, University of California, Santa Cruz.
Orgun, C. O. and Sprouse, R. L. 1999. From MPARSE to CONTROL: deriving ungrammaticality. Phonology 16:191-224.

“Python” is a proper name

In just the last few days I’ve seen a half dozen instances of the phrase python package or python script in published academic work. It’s disappointing to me that this got by the reviewers, action editors, and copy editors, since Python is obviously a proper name and should be in titlecase. (The fact that the interpreter command is python is irrelevant.)

“…phonology is logically (and causally) prior to phonetics.”

Two important consequences follow from this. First, that phonology is logically (and causally) prior to phonetics as here defined. Second, phonology is also epistemologically prior to phonetics. Judgments about phonetic events are invariably made in terms of perceptual phonology. (Hammarberg 1976:356)

In this post I’d like to briefly review a view of the relationship between phonetics and phonology as related by Hammarberg (1976) and Appelbaum (1996), the former being primarily concerned with production and the latter with perception.

Phonetics, being concerned with the material and physical, has tended to align itself with the physical sciences (and physics in particular), and with the empiricist tradition in science.1,2 In contrast, much of what has been called the cognitive revolution in the cognitive sciences, and in linguistics in particular, is explicitly anti-empiricist. As Hammarberg and Appelbaum argue, the empiricist biases of phonetics make it ill-suited to explain fundamental facts about speech.

It is generally understood that spoken language is not produced as a discrete sequence but rather a series of overlapping gestures and acoustic signatures. Anyone who has looked closely at the acoustics of speech will already recognize that it is impossible to say exactly where, in a word like cat, the [æ]-ness ends and the [t]-ness begins. In a worrd like soon, the fricative portion shows signs of rounding not found in words like scene. From an acoustic record alone, one cannot determine empirically how many segments are present. And, one cannot produce natural-sounding synthesized speech via simple concatenation of segments. It is not just that the [æ, t, s] and other segments are coarticulated with nearby segments, however: it is also the case that there are simply no invariant acoustic-phonetic properties that uniquely characterize [t]. A [t] spoken by a child, by a man with a mouth full of chili, by a woman missing her front teeth, and so on may have radically different acoustic properties, yet we as scientists understand them to be in some sense identical phenomena.

This is a basic principle of scientific discovery: one must assume that “the vast multitude of phenomena he encounters may be accounted for in terms of the interactions of a fairly small number of basic entities, standard elementary individuals. His task thus becomes one of identifying the basic entities and describing the interactions in virtue of which the encountered phenomena are generated. From this emerge our…notions of the identity and nonidentity of phenomena.” (Hammarberg, p. 354) The linguistic notion of segment is perhaps the most important of these basic entities. It is an entity recognized both by those early lay-linguists, the Iron Age scribes who gave us the alphabet, as well as one of the most venerable notions in the history of modern linguistics. Yet, segments do not have a physical reality of their own; they do not exist in the physical world, but only in the human mind. They are “internally generated, the creature of some kind of perceptual-cognitive process.”

It is generally uncontroversial to speak of the output of the phonological component as the input to the phonetic component. From this it follows that phonology is cognitively and epistemically prior to phonetics. Coarticulation, for instance, results because of the process which maps segments—which, remember, exist only in the mind of speakers—onto articulatory and acoustic events. But one cannot talk about coarticulation without segments, since it is the spreading of articulatory-acoustic properties between segments that defines coarticulation. One must know that /s/ exists, and has an inherent properties not normally associated with—or compatible with—lip rounding to even observe the anticipatory lip rounding in words like soon.

The existence of coarticulation is often understood teleologically, in the sense that is taken to be in part mechanical, automatic, inertial. This too is a mistake, according to Hammarberg: apparent teleological explanations of human behavior should be recast, as is the tradition in Western philosophy, as the result of intentional, causal behavior. The existence of anticipatory articulation shows us that the influence of the /u/ in soon has on the realization of the preceding /s/ occurred some time before instructions to the articulators were generated, and the level at which this influence occurs should therefore be identified with the mental rather than the physiological. Hammarberg continues to argue that coarticulatory processes are akin to ordinary allophony and should reside in the scope of phonological theory. This argument is strengthened insofar as coarticulation has a language-specific character, as is sometimes claimed.

Appelbaum, while not citing Hammarberg’s original paper, extends this critique to the theory of speech perception. It is an assumption of the so-called motor theory that there are invariant properties which identify “phonetic gestures”. Since the motor theorists do not present any evidence that such invariants soc much as exist, we instead must be abstract out into mental entities which have all the properties of—and which Appelbaum identifies with—what we are calling segments, or perhaps lower-level entities like phonological features. Under this approach, then, there is no content to the motor theory of speech perception beyond the obvious point that phonetic experience, somehow, turns into purely mental representations. Again, the empiricist biases of phonetics have lead us astray.

The above discussion may influence the way we think about the role of phonetics in linguistics education. Phonetics is generally viewed as its own autonomous subdiscipline, and modern acoustic and articulatory analysis is certainly complex enough to justify serious graduate instruction, but it would seem to suggest that phonetic tools exist primarily as a way of gathering phonological information rather than  as an autonomous discipline. I am not sure I am ready to conclude that, but it certainly is provocative!

Endnotes

  1. Empiricism refers to a theory of epistemology and should not be confused with the empirical method in science (the use of sense-based observation). Many prominent thinkers reject empiricism in favor of rationalism, but support the use  of empirical methods. No one is seriously arguing against the use of the senses.
  2. This will be shown to be yet another example of physics envy as the source of sloppy thinking in linguistics.

References

Appelbaum, I. 1996. The lack of invariance problem and the goal of speech perception. In Proceeding of Fourth International Conference on Spoken Language Processing, 1541-1544.
Hammarberg, R. 1976. The metaphysics of coarticulation. Journal of Phonetics 4: 353-363.