Comments

Showing posts with label combinatorics. Show all posts
Showing posts with label combinatorics. Show all posts

Thursday, July 18, 2013

Why Morphology?


For quite a while now, I’ve been wondering why natural language (NL) has so much morphology.  In fact, if one thinks about morphology with a minimalist mind set one gets to feeling a little like I. I. Rabi did regarding the discovery of the muon. His reaction? “Who ordered that?”.  So too with morphology; what’s it doing and why is there both so much of it in some NLs (Amer-Indian languages) and so little of it in others (Chinese, English)? One thing seems certain, look around and you can hardly miss the fact that this is a characteristic feature of NLs in spades!!

So what’s so puzzling? Two things. First, it’s absent from artificial languages, in contrast to, say, unbounded hierarchy and long distance dependency (think operator-variable binding). Second, it’s not obviously functionally necessary (say to facilitate comprehension). For example, there is no obvious reason to think that Chinese or English speakers (there is comparatively little morphology here) have more difficulty communicating with one another than do Georgian or Athabaskan speakers, despite the comparative dearth of apparent morphology. In sum, morphology does not appear to be conceptually or functionally necessary for otherwise we (I?) might have expected it to be even more prevalent than it is. After all if it’s really critical and/or functionally useful then one might expect it to be everywhere, even in our constructed artificial languages.  Nonetheless, it’s pretty clear that NLs hardly shy away from morphological complexity.

Moreover, it appears that kids have relatively little problem tracking it. I have been told that whereas LADs (language acquisition devices, aka: kids) omit morphology in the early stages of acquisition (e.g. ‘He go’), they don’t produce illicit “positive” combinations (e.g. ‘They leaves’). I have even been told that this holds true for languages with rich determiner systems and noun classes and fancy intricate verbal morphology: it seems that kids are very good at correctly classifying these and quickly master the relevant morphological paradigms.  So, LADs (and LASs; Language Acquisition Systems) are good at learning these horrifying (to an outsider or second language learner) details and at deploying them effectively as native speakers. So, again, why morphology?

Unburdened by any knowledge of the subject matter, I can think of four possible reasons for morphology’s ubiquity within NLs. I should add that what follows is entirely speculative and I hope that this post motivates others to speculate as well. I would love to have some ideas to chase down. So here goes.

The first possibility is that visible morphology is a surface manifestation of a deeper underlying morphology. This is a pretty standard Generative assumption going back to the heyday of comparative syntax in the early 80s. The first version of this was Jean-Roger Vergnaud’s (terrific) theory of abstract case. The key idea is that all languages have an underlying abstract case system that regulates the distribution of nominal expressions.  If we further assume that this abstract system can be phonetically externalized, then the seeds of visible morphology are inherent in the fundamental structure of FL. The general principle then is that abstract morphemes (provided by UG) are wont to find phonetic expression (are mapped to the sensory and motor systems (S&M)), at least some of the time.

This idea has been developed repeatedly. In fact, the following is still a not an unheard of move: We find property P in grammar G overtly, we then assume that something similar occurs in all Gs, at least covertly. This move is particularly reasonable in the context of “Greed” based grammars characteristic of early minimalism. If all operations are “forced” and the force reduces to checking abstract features, then using the logic of abstract case theory, we should not be surprised if a GL expresses these phonetically.

Note that if something like this is correct (observe the if), then the existence of overt morphology is not particularly surprising, though the question remains of why some Gs externalize these abstracta and some remain phonetically more mum.  Of late, however, this Greed based approach has dimmed somewhat (or at least that’s my impression) and generate and filter models of various kinds are again being actively pursued. So…

A second way to try and explain morphology piggy-backs on Chomsky’s recent claims that Gs are not pairings of sound and meaning but pairings of meanings with sound. His general idea is that whereas the mapping from lexical selection to CI is neat and pretty, the externalization to the S&M systems is less straightforward. This comports with the view that the first real payoff to the emergence of grammar was not an enhancement of communication but a conceptual boost expanding the range of cognitive computations in the individual, i.e. thinking and planning (see here). Thus externalization via S&M is a late add-on to an already developed system. This “extra” might have required some tinkering to allow it to hook onto the main lexicon-to-CI system and that tinkering is manifest as morphology. In effect then, morphology is family related to Chomsky and Halle’s old readjustment rules.  From where I sit, some of the work in Distributed Morphology might be understood in this way (it packages the syntax in ways palpable to S&M), though, I am really no expert in these matters so beware anything I say about the topic. At any rate, this could be a second source for morphology, a kluge to get Gs to “talk.”

I can think of a third reason for overt morphology that is at right angles to these sorts of more grammatically based considerations. There are really two big facts about human linguistic facility: (i) the presence of unbounded hierarchical Gs and (ii) the huge vocabulary speakers have.  Though it’s nice to be able to embed, it’s also nice to have lots of words.  Indeed, if travelling to a foreign venue where residents speak V and given the choice of 25 words of V plus all of GV or 25,000 words of V plus just the grammar of simple declaratives (and maybe some questions), I’d choose the second over the first hands down. You can get a lot of distance on a crappy grammar (even no grammar) and a large vocabulary.  So, here’s the thought: might morphology facilitate vocabulary development?  Building a lexicon is tough (and important) and we do it rapidly, very rapidly. Might overt morphology aid this process, especially if word order in a given language (and hence PLD of that language) is not all that rigid?  It could aid this process by providing stable landmarks near which content words could be found. If transitional probabilities are a tool for breaking into language (and the speech stream, as Chomsky proposed in LSLT and later rediscovered by Aislin, Saffran and Newport), then having morphological landmarks that probabilistically vary at different rates than the expressions that sit within these landmarks then it might serve to focus LADs and LASs on the stuff that needs learning; content words. On this story, morphology exists to make word learning easier by providing frames within a sentence for the all-important lexical content material.

There is a second version of this kind of story that I would like to end with. I should warn you that it is a little involved. Here goes. Chomsky has long identified two surprising properties of NLs. The first is unbounded hierarchical recursion, the second is our lexical profligacy. We not only can combine words but we have lots of words to combine. A typical vocabulary is in the 50,000 word range (depending on how one counts). How do we do this. Well, assume that at the very least, each new vocabulary item consists of some kind of tag (i.e. a sound or a hand gesture). In fact, for simplicity say that acquiring a word is simply tagging it (this is Quin’s “museum myth,” which like many myths may in fact be true). Now this sounds like it should be fairly easy, but is it?  Consider manufacturing 50,000 semantically arbitrary tags (remember, words don’t sound the way they do because they mean what the do, or vice versa).  This is hard. To do this effectively requires a combinatoric system, Indeed, something very like a phonology, which is able to combine atomic units into lexical complexes. So, assume that to have a large lexicon we need something like a combinatoric phonology and the products of this system are the atoms that the syntax combines into further hierarchically structured complexes. Here’s the idea: morphology mediates the interactions of these two very different combinatoric systems.  Meshing word structures and sentence structure is hard because the modes of combination of the two kinds of systems are different. Both kinds play crucial (and distinctive) roles in NL and when they combine morphology happens!  So, on this conception, morphology is not for lexical acquisition, but exists to allow words with their structures to combine into phrases with their structures.

The four speculations above are, to repeat, all very speculative and very inchoate. They don’t appear to be mutually inconsistent, but this may be because they are so lightly sketched. The stories are most likely naïve, especially so given my virtually complete ignorance of morphology and its intricacies. I invite those of you who know something about morphology to weigh in. I’d love to have even a cursory answer to the question.