Comments

Showing posts with label Matzel. Show all posts
Showing posts with label Matzel. Show all posts

Monday, October 7, 2013

Domain Specificity


One of the things that seems to bug many about FL/UG is the supposition that it is a domain specific module dedicated to ferreting out specifically linguistic information. Even those that have been reconciled to the possibility that minds/brains come chock full of pre-packaging, seem loath to assume that these natively provided innards are linguistically dedicated.  This antipathy has even come to afflict generativists of the MP stripe for there is believed to be an inconsistency between domain specificity and the minimalist ambition to simplify UG by removing large parts of its linguistically idiosyncratic structure. I have suggested elsewhere (here) that this tension is only apparent, and that it is entirely possible to pursue the MP cognitive leveling strategy without abandoning the idea that there is a dedicated FL/UG as part of human biological endowment. In a new paper, Gallistel and Matzel (G&M) (here) argue that domain specificity is the biological default once one rejects associationism and adopts an information processing model of cognition. Put more crudely, allergy to domain specificity is just another symptom of latent empiricism (i.e. a sad legacy of associationism).

I, of course, endorse G&M’s position that there is nothing inconsistent between accepting functionally differentiated modules and the assumption that these are largely constructed using common basic operations. And, I of course love G&M’s position that once one drops any associationist sympathies (and I urge you all to immediately do this for your own intellectual well being!), then the hunt for general learning mechanisms looks, at least in biological domains, ill-advised.  Or put more positively: once one adopts an information processing perspective then domain specificity seems obviously correct. Let’s consider G&M’s points in a little detail.

G&M contrasts associationist (A) and information processing (IP) models of learning and memory. The G&M paper is divided into two parts, more or less. The first several pages comprise a concise critique of associationist/neural net models in which learning is “the rewiring of a plastic nervous system by experience, and memory resides in the changed wiring (170).” The second part develops the evidence for an IP perspective on neural computations. The IP models contrast with A-models in distinguishing the mechanisms for learning (whose function is to “extract potentially useful information from experience”) and those for memory (whose function is to “carr[y] the acquired information forward in time in a computationally accessible form that is acted upon by the animal at the time of retrieval”) (170).  Here are some of their central points.

A-models are “recapitulative.” What G&M (170) intend here is that learning consists in finding the pattern in the data (see here): “An input that is part of the training input, or similar to it, evokes the trained output, or an output similar to it.”  IP models are “in no way recapitulations of the mappings (if any) that occurred during the learning.” This is the classical difference between rationalist vs empiricist conceptions of learning. A- models conceive of environmental input as adapting “behavior” to environmental circumstances. IP-models conceive of learning as building a “representation of important aspects of the experienced world.”

A-models gain a lot of their purchase within neuro-science (and psychology) by appearing to link so directly to a possible neural mechanism; long-term potentiation (LTP). However, G&M argue vigorously that LTP support for A-models entirely evaporates when the evidence linking LTP to A-models is carefully evaluated.  G&M walk us slowly through the various disconnects between standard A-processes of learning and LTP function; their time scales are completely different (“…temporal properties of LTP do not explain the temporal properties of behaviorally measured association formation (172).”), their persistence (i.e. how long the changes in LTP vs associations last) is entirely different and so “LTP does not explain the persistence of associative learning (172),” their reactivation schedules are entirely different (i.e. If L is learned and then extinguished, L is reacquired more quickly, but “LTP is neither more easily nor more persistent than it was after previous inductions.”), nor do LTP models provide any mechanism for solving the encoding problem (viz. A-learning is mediated by the comparison of different kinds of temporal intervals and there is no obvious way for LTP nets to do this) except by noting that what gets encoded is emergent (and this amounts to punting on the encoding problem, rather than addressing it).

In short, there is no support for A-models from neural LTP models. Indeed, the latter seem entirely out of synch with what’s needed to explain memory and learning. As G&M put it: “…if synaptic LTP is the mechanism of associative learning- and more generally, of memory- then it is disappointing that its properties explain neither the basic properties of associative learning nor the essential properties of a memory mechanism (173).” So much for the oft insinuated claim that connectionist models are preferable because they are neurally plausible (indeed, obvious!).

General conclusion: A-models have no obvious support from standard LTP models and these standard LTP models are inadequate for handling the simplest behavioral data. In effect, A- and LTP- accounts are the wrong kinds of theories (not wrong in detail, but in conception and hence without much (if any) redeeming scientific value) if one is interested in understanding the neural bases of cognition.

So what’s the right approach? IP-models. In the last parts of the paper G&M go over some examples.  They note that biologically plausible IP models will all share some important features:

1.     They will involve domain specific computations. Why? “Because no general purpose computation could serve the demands of all types of learning (175),” i.e. domain specificity is the natural expectation for IP models of neuro-cognition.
2.     The different computations will apply the same “primitive operations” in achieving functionally different results (175).[1]
3.     The IP approach to learning mechanisms “requires an understanding of the rudiments of the different domains in which the different learning mechanisms operate” (175). So, for example, figuring out if A is cause of B, or A is the edge of B will involve different computations from each other and from those that mediate the pairing of meanings and/with sounds.
4.     Though the neuro-science is at a primitive stage right now, “…if learning is the result of domain-specific computations, then studying the mechanism of learning is indistinguishable from studying the neural mechanisms that implement computations (175).”

Note that this will hold as much in the domain of language as in navigation and spatial representation.  In other words, once one dumps Associationism (as one must as it is empirically completely inadequate and intellectually toxic) then domain specificity is virtually ineluctable. There exist no interesting general purpose learning systems (just as there it no general sensing mechanism, as Gallistel has been wont to observe). That’s the G&M message. Cognitive computation, if it’s to be neurally based, will be quite specifically tailored to the cognitive tasks at hand, even if built from common primitive circuits.

The most interesting part of G&M, at least to me, was the review of the specific neural cells implicated in animal capacities for locating oneself in space and moving around within it. It seems that neuro-scientists are finding “functionally specialized neurons [that] signal abstract properties of the animal’s relation to its spatial environment (185).” These are genetically controlled and, as G&M note, their functional specialization provide “compelling evidence for problem-specific mechanisms.”

Note that the points G&M make above fit very snugly with standard assumptions within the Chomsky version of the generative tradition. In other words, the assumptions that generative linguists make concerning domain specific computations and mechanisms (though not necessarily primitive operations) simply reflects what is, or at least should be, the standard assumption in the study of cognition once Associationism is dumped (may it’s baneful influence soon disappear). If G&M are right, then there are no good reasons from neuro-biology for thinking that the standard assumptions concerning native domain specific structures for language are exotic or untoward.  They are neither. The problem is not with these assumptions, but with the unholy alliance between some parts of contemporary neuroscience and the A-models of learning and cognition that neuro types have uncritically accepted.

If you read the whole G&M paper (some parts involve some heavy lifting) and translate it into a linguistics framework, it is very hard to avoid the conclusion that if G&M are correct, (and, in case you’ve missed it, IMO they are) then the Chomskyan conception of language, mind, and brain is both anodyne and the only plausible game in cog-neuro town.


[1] A similar point wrt MP and UG is made here.