Comments

Tuesday, April 18, 2017

Inference to some great provocations

David Berlinsky is editing a newish online magazine Inference that articles in which I have mentioned in several previous posts. The latest issue is full of fun for linguists as there are four articles of immediate relevance. Here’s the link for the issue. Let me say a word or two about the pieces.

The first is an essay by Chomsky that goes over familiar ground regarding the distinctive nature of human linguistic capacity. He observes that this observation has a Cartesian pedigree and that language was recognized as distinctive (all and only humans have it) and wondrous (it was free and capable of expressing unboundedly many thoughts) and demanding of some kind of explanation (it really didn’t fit in well with what was understood to be the causal structure of the physical world) as early as it was noticed.

As Chomsky notes, Cartesians had relatively little of substance to say about the underpinnings of this wondrous capacity, mainly because the 17th century lacked the mathematical tools for the project. They had no way of describing how it was possible to “make infinite use of finite means” as von Humboldt put it (2). This changed in the 20th century with Church, Godel, Post and Turing laying the foundations of computation theory. This work “demonstrated how a finite object like the brain could generate an infinite variety of expressions.” And as a result, “[i]t became possible, for the first time, to address part of” the problem that the Cartesians identified (2).

Note the ‘part of’ hedge. As Chomsky emphasizes, the problem the Cartesians identified has two parts. The first, and for them the most important feature, is the distinction between “inclined” vs “impelled” behavior (3). Machines are impelled to act, never “inclined.” Humans, being free agents, are most often “inclined” (though they can be “compelled” as well). Use of language is the poster child for inclined behavior. Cartesians had no good understanding of the mechanics of inclination. As Chomsky observes, more than 300 years later, neither do we. As he puts it, language’s “free creative use remains a mystery,” as does free action in general (e.g. raising one’s hand) (3).

The second part, one that computation theory has given us a modest handle on, is the unbounded nature of the thoughts we can express. This feature very much impressed Galileo and Arnauld & Lancelot and von Humboldt, and it should impress you too! The “infinite variety” of meaningfully distinct expressions characteristic of human language “surpasse[s] all stupendous inventions” (1).  Chomsky has redubbed this feature of language “the Basic Property” (BP). BP refers to a property of the human brain, “the language faculty,” and its capacity to “construct a digitally infinite array of structured expressions” each of which “is semantically interpreted as expressing a thought, and each can be externalized by some sensory modality such as speech” (2).  BP is what GG has been investigating for the last 60 years or so. Quite a lot has been discovered about it (and yes, there is still lots that we don’t know!).

Chomsky emphasizes something that is worth reemphasizing: these facts about language are not news. That humans have linguistic creativity in the two senses above should not really be a matter of dispute. That humans do language like no other animal does should also be uncontroversial. How we do this is a very tough question, only a small part (very small part) of which we have managed to illuminate. It is sad that much debate still circulates around the whether question rather than the how. It is wasted time.

An important theme in Chomsky’s essay turns on how the world looks when we have no idea what’s up. Here is a quote that I believe all good scientifically inclined GGers should have tattooed to themselves (preferably in some discrete place) (3):

When understanding is thin, we expect to see extreme variety and complexity.

Absolutely! Variety and complexity are hallmarks of ignorance. And this is why progress and simplicity go hand in hand. And this is why I have clasped to my heart Dresher’s apposite dictum: There should be only two kinds of papers in linguistics: (i) papers that show that two things that look completely different are roughly the same and (ii) papers that show that two things that are roughly the same are in fact identical. These are the papers that highlight our progressively deeper understanding. Complication is often necessary, but it is progressive just in case it paves the way for greater simplicity.

The unification and simplicity is, thus, a leading indicator of scientific insight. Within linguistics it has a second function. It allows one to start addressing the issue of how FL might have evolved.  Here’s Chomsky:

In the analysis of the Basic Property, we are bound to seek the simplest computational procedure consistent with the data of language. Simplicity is implicit in the basic goals of scientific inquiry. It has long been recognized that only simple theories can attain a rich explanatory depth. “Nature never doth that by many things, which may be done by a few,” Galileo remarked, and this maxim has guided the sciences since their modern origins. It is the task of the scientist to demonstrate this, from the motion of the planets, to an eagle’s flight, to the inner workings of a cell, to the growth of language in the mind of a child. Linguistics seeks the simplest theory for an additional reason: it must face the problem of evolvability. Not a great deal is known about the evolution of modern humans. The few facts that are well established, and others that have recently been coming to light, are rather suggestive. They conform to the conclusion that the language faculty is very simple; it may, perhaps, even be computationally optimal, precisely what is suggested on methodological grounds.

Unless FL is simpler than we have considered it to be up till now (e.g. far simpler than say GBish models make it out to be) then there is little chance that we will be able to explain its etiology. So there are both general methodological grounds for wanting simple theories of FL and linguistic internal reasons for hoping that much of the apparent complexity of FL is just apparent.

Chomsky’s piece proceeds by rehearsing in short form the basic minimalist trope concerning evolvability. First, that we know little about it and that we will likely not know very much about it ever. Second, that FL is a true species property as the Cartesians surmised. Third, that FL has not evolved much since humans separated. Fourth, that FL is a pretty recent biological innovation. The third and fourth points are taken to imply that the Basic Property aspect of FL must be pretty simple in the sense that what we see today pretty well reflects the original evo innovation and so its properties are physically simple in that they have not been shaped by the forces of selection. In other words, what we see in BP is pretty much undistorted by the shaping effects of evolution and so largely reflect the physical constraints that allowed it to emerge.

All of this is by now pretty standard stuff, but Chomsky tells it well here. He goes on to do what any such story requires. He tries to illustrate how a simple system of the kind he envisions will have those features that GG has discovered to be characteristic of FL (e.g. structure dependence, unboundedly many discrete structures capable of supporting semantic interpretation etc.). This second step is what makes MP really interesting. We have a pretty good idea what kinds of things FL concerns itself with. That’s what 60 years of GG research has provided. MP’s goal is to show how to derive these properties from simpler starting points, the simpler the better. The target of explanation (the explanadum) are the “laws” of GB. MP theories are interesting to the degree that they can derive these “laws” from simpler more principled starting points. And, that, Chomsky argues, is what what makes Merge based accounts interesting, they derive features that we have every reason to believe characterize FL.[1]

Two other papers in the issue address these minimalist themes. The first is a review of the recent Berwick & Chomsky (B&C) book Why only us. The second is a review of a book on the origins of symbolic artifacts. Cederic Boeckx (CB) reviews B&C. Ian Tatersall (IT) reviews the second. The reviews are in interesting conflict.

The Boeckx review is quite negative, the heart of the criticism being that asking ‘why only humans have language’ is the wrong question. What makes it wrong? Well, frankly, I am not sure. But I think that the CB review thinks that asking it endorses a form of “exceptional nativism” (7) that fails to recognize “the mosaic character of language,” which, if I get the point, implies eschewing “descent with modification” models of evolution (the gold standard according to CB) in favor of “top-down, all-or-nothing” perspectives that reject comparative cognition models (or any animal models), dismiss cultural transmission as playing any role in explaining “linguistic complexity” and generally take a jaundiced view of any evolutionary accounts of language (7-8). I actually am skeptical regarding any of this.

Before addressing these points, however, it is interesting that IT appears to take the position that CB finds wrong-headed. He thinks that human symbolic capacities are biologically quite distinctive (indeed “unique”) and very much in need of some explanation. Moreover, in contrast to CB, IT thinks it pretty clear that this “symbolic activity” is of “rather recent origin” and that, “as far as can be told, it was only our lineage that achieved symbolic intelligence with all of its (unintended) consequences” (1). If we read “symbolic” here to mean “linguistic” (which I think is a fair reading), it appears that IT is asking for exactly the kind of inquiry that CB thinks misconceived.

That said, let’s return to CB’s worries. The review makes several worthwhile points. IMO, the two most useful are the observation that there is more to language evolution than the emergence of the Basic Property (i.e. Merge and discretely infinite hierarchically structured objects) and that there may be more time available for selection to work its magic than is presupposed.  Let’s consider these points in turn.

I think that many would be happy to agree that though BP is a distinctive property of human language it may not be the only distinctive linguistic property. CB is right to observe that if there are others (sometimes grouped together as FLW vs FLN) then these need to be biologically fixed and that, to date, MP has had little to say about these. One might go further; to date it is not clear that we have identified many properties of FLW at all. Are there any?

One plausible candidate involves those faculties recruited for externalization. It is reasonable to think that once FLN was fixed in the species, that linking its products to the AP interface required some (possibly extensive) distinctive biological retrofitting. Indeed, one might imagine that all of phonology is such a biological kludge and that human phonology has no close biological analogues outside of humans.[2] If this is so, then the question of how much time this retrofitting required and how fast the mechanisms of evolution (e.g. selection) operate is an important one. Indeed, if there was special retrofitting for FLW linguistic properties then these must have all taken place before the time that humans went their separate ways for precisely the reasons that Chomsky likes to (rightly) emphasize: not only can any human acquire the recursive properties of any G, s/he can also acquire the FLW properties of any G (e.g. any phonology, morphology, metrical system etc.).[3] If acquiring any of these requires a special distinctive biology, then this must have been fixed before we went our separate ways or we would expect, contrary to apparent fact, that e.g. some “accents” would be inaccessible to some kids. CB is quite right that it behooves us to start identifying distinctive linguistic properties beyond the Basic Property and asking how they might have become fixed. And CB is also right that this is a domain in which comparative cognition/biology would be very useful (and has already been started (see note 2). It is less clear that any of this applies to explaining the evolution of the Basic Property itself.

If this is right, it is hard for me to understand CB’s criticism of B&C’s identification of hierarchical recursion as a very central distinctive feature of FL and asking how it could have emerged.  CB seems to accept this point at times (“such a property unquestionably exists” (3)) but thinks that B&C are too obsessed with it. But this seems to me an odd criticism. Why? Because B&C’s way into the ling-evo issues is exactly the right way to study the evolution of any trait: First identify the trait of interest. Second, explain how it could have emerged.  B&C identify the trait (viz. hierarchical recursion) and explain that it arose via the one time (non-gradual) emergence of a recursive operation like Merge. The problem with lots of evo of lang work is that it fails to take the first step of identifying the trait at issue. But absent this any further evolutionary speculation is idle. If one concedes that a basic feature of FL is the Basic Property, then obsessing about how it could have emerged is exactly the right way to proceed.

Furthermore, and here I think that CB’s discussion is off the mark, it seems pretty clear that this property is not going to be all that amenable to any thing but a “top-down, all-or-nothing” account. What I mean is that recursion is not something that takes place in steps, a point that Dawkins made succinctly in support of Chomsky’s proposal (see here). As he notes, there is no such thing as “half recursion” and so there will be no very interesting “descent with modification” account of this property. Something special happened in humans. Among other things this led to hierarchical recursion. And this thing, whatever it was, likely came in one fell swoop. This might not be all there is to say about language, but this is one big thing about it and I don’t see why CB is resistant to this point. Or, put another way, even if CB is right about many other features of language being distinctive and amenable to more conventional evo analysis, it does not gainsay the fact that the Basic Property is not one of these.

There is actually a more exorbitant possibility that perhaps CB is reacting to. As the review notes (7): “Language is special, but not all that special; all creatures have special abilities.” I don’t want to over-read this, but one way of taking it is that different “abilities” supervene on common capacities. This amounts to a warning not to confuse apparent expressions of capacities for fundamental differences in capacities. This is a version of the standard continuity thesis (that Lenneberg, among others, argued is very misleading (i.e. false) wrt language). On this view, there is nothing much different in the capacities of the “language ready” brain from the “language capable” brain. They are the same thing. In effect, we need add nothing to an ape brain to get ours, though some reorganization might be required (i.e no new circuits). I personally don’t think this is so. Why? For the traditional reasons that Chomsky and IT note, namely that nothing else looks like it does language like we do, even remotely. And though I doubt that hierarchical recursion is the whole story (and have even suggested that something other than Merge is the secret sauce that got things going), I do think that it is a big part of it and that downplaying its distinctiveness is not useful.

Let me put this another way. All can agree that evolution involves descent with modification. The question is how big a role to attribute to descent and how much to modification (as well as how much modification is permitted). The MP idea can be seen as saying that much of FL is there before Merge got added. Merge is the “modification” all else the “descent.” There will fe features of FL continuous with what came before and some not continuous. No mystery about the outline of such an analysis, though the details can be very hard to develop. At any rate, it is hard for me to see what would go wrong if one assumed that Merge (like the third color neuron involved in trichromatic vision (thx Bill for this)) is a novel circuit and that FL does what it does by combining the powers of this new operation with those cognitive/computational powers inherited from our ancestors. That would be descent with modification. And, so far as I can tell, that is what a standard MP story like that in B&C aims to deliver. Why CB doesn’t like (or doesn’t appear to like) this kind of story escapes me.

Observe that how one falls on the distinctiveness of BC issue relates to what one thinks of the short time span observation (i.e. language is of recent vintage so there is little time for natural selection or descent with modification to work its magic). The view Chomsky (and Berwick and Dawkins and Tatersall) favor is that there is something qualitatively different between language capable brains and ones that are not. This does not mean that they don’t also greatly overlap. It just means that they are not capacity congruent. But if there is a qualitative difference (e.g. a novel kind of circuit) then the emphasis will be on the modifications, not the descent in accounting for the distinctiveness. B&C is happy enough with the idea that FL properties are largely shared with our ancestors. But there is something different, and that difference is a big deal. And we have a pretty good idea about (some of) the fine structure of that difference and that is what Minimalist linguistics should aim to explain.[4] Indeed, I have argued and would continue to argue that the name of the Minimalist game is to explain these very properties in a simple way. But I’ve said that already here, so I won’t belabor the point (though I encourage you to do so).

A few more random remarks and I am done. The IT piece provides a quick overview of how distinctive human symbolic (linguistic?) capacities are. In IT’s view, very. In IT’s view, the difference also emerged very recently, and understanding that is critical to understanding modern humans. And he is not alone. The reviewee Genevieve von Petziger appears to take a similar view, dating the start of the modern human mind to about 80kya (2). All this fits in with the dates that Chomsky generally assumes. It is nice to see that (some) people expert in this area find these datings and the idea that the capacity of interest is unique to us credible. Of course, to the degree that this dating is credible and to the degree that this is not a long time for evolution to exercise its powers the harder the evolutionary problem becomes. And, of course, that’s what makes the problem interesting. At any rate, what the IT review makes clear is that the way Chomsky has framed the problem is not without reasonable expert support. Whether this view is correct, is, of course, an empirical matter (and hence beyond my domain to competently judge).

Ok, let me mention two more intellectual confections of interest and we are done. I will be short.

The first is a review of Wolfe’s book by David Lobina and Mark Brenchley. It is really good and I cannot recommend it highly enough. I urge you in particular to read the discussion on recursion as self-reference vs self-embedding and the very enlightening discussion of how Post’s original formalism (might have) led to some confusion on these issues. I particularly liked the discussion of how Merge de-confuses them, in effect by dumping the string based conception of recursion that Post’s formalism used (and which invited a view of recursion as self-embedding) and implementing the recursive idea more cleanly in a Merge like system in which linguistic structures are directly embedded in one another without transiting through strings at all. This cleanly distinguishes the (misleading) idea that the recursion lies with embedding clauses within clauses from the more fundamental idea that recursion requires some kind of inductive self-reference. Like I said, the discussion is terrific and very useful.

And now for desert: read David Adger’s fun review of Arrival. I confess that I did not really like the movie that much, but after reading David’s review, I intend to re-see it with a more open mind.

That’s it. Take a look at the issue of Inference. It’s nice to see serious linguistic issues intelligently discussed in a non-specialist’s venue. It can be done and done well. We need more of it.




[1] Chomsky also mentions that how lexical items have very distinctive properties and that we understand very little about them. This ahs become a standard trope in his essays, and a welcome one. It seems that lexical items are unlike animal signs in that the latter are really “referential” in ways that the former are not. The how and whys behind this, however, is completely opaque.
[2] There has been quite a lot of interesting comparative work done, most prominently by Berwick, on relating human phonology with bird song. See here and here for some discussion and links.
[3] There is another possibility: once FLN is in place there is only one way to retrofit all the components of FLW. If so, then there is no selection going on here and so the fact that all those endowed with FLNs share common FLWs would not require a common ancestor for the FLWs. Though I know nothing about these things, this option strikes me as far-fetched. If it is, then the logic that Chomsky has deployed for arguing that FLN was in place before humans went their separate ways would hold for FLW as well.
[4] CB makes a claim that is often mooted in discussions about biology. It is Dobzhansky’s dictum that nothing in biology makes sense except in the light of evolution. I think that this is overstated. Lots of biology “makes sense” without worrying about origin. We can understand how hearts work or eyes see or vocal tracts produce sounds without knowing anything at all about how they emerged. This is not to diss the inquiry: we all want to know how things came to be what they are. But the idea that natural selection is the only thing that makes sense of what we see is often overblown, especially so when Dobzhansky quotes are marshaled. For some interesting discussion of this see this.

10 comments:

  1. >There should be only two kinds of papers in linguistics: (i) papers that show that two things that look completely different are roughly the same and (ii) papers that show that two things that are roughly the same are in fact identical.

    So all papers about E-languages are useless and should not exist? The entirety of the Labovian enterprise, say? Also the entirety of historical linguistics? Documentation and description of endangered languages is right out, I guess, given how it focus on surface, language-specific properties (a kind of data which, unlike the general faculty of language, will likely disappear within a generation)? No more papers on forensic linguistics? Natural language processing, data extraction from textual corpora, the cognition of reading and writing? Methods to help post-critical-period adults learn new languages better and faster? Everyone who studies these things should be defunded ASAP, their journals should go under etc.?

    Or are we allowed to investigate these things, only we may not use the elevated term "linguistics" for them? "Linguistics" is reserved for the study of the biological faculty of language, and every other scientific investigation of language should be called some other, lesser thing?

    ReplyDelete
    Replies
    1. You can investigate whatever you want. I am, sadly, not Tzar of what gets done in the field. My point was more modest: to the degree you are interested in the structure of FL (which I take to be the goal of work in GG) then these are the kinds of papers you are interested in as these are the kinds of papers that encourage explanation. Other work is interesting to the degree that it fosters this kind. So, yes, if you care to explain this is the kind of paper you will write. Of course, explanation isn't the only thing in life nor do I control funding/publication etc (nor would I want to despite my snide remark above). What I DO think is that for my concerns (which I admit are not widely shared) Dresher's point has hit a worthy nail right on the head.

      Delete
  2. Just a short point about L&B on recursion... It is s nice piece, and I think they are basically right, but there is a deep connection between recursion and self-embedding that they downplay. A recursive df. enumerates a set, where the members don't have to be specified as involving self-embedding. That's right, but such members understood as values of the function do involve embedding of the self-referred to concept qua enumerated. So, to use the arithmetic cases, instead of thinking of 1 and 2, say, as distinct objects, we think of them as s(0) and s(s(0)), respectively; mutates mutandis for the factorial, where 6, say, is (((1)x2)x3). Lord knows Wolfe knows shit, but the ideas are tricky.

    ReplyDelete
    Replies
    1. Great points--that was one of the problems I had with L&B's piece (which I didn't find all that helpful, to be honest). I think it would've been stronger if they'd drawn a clearer distinction between recursion and self-embedding, because their examples of recursion do involve embedding; of *course* factorial 4 has factorial 3, 2, and 1 embedded within. The decision to end on the Fitch quote seemed quite strange (I thought Fitch was there to look for recursion, not embedding).

      Delete
    2. Thanks, Steven. Yes, I thought the distinction remains a bit unclear in the paper. Also, yes, not sure what the Fitch quote means. Is it that he was after evidence for embedding because that is what DE had erroneously conflated with recursion?

      Delete
  3. My review of B&C raises a number of quite different objections, mainly focused on chapter 4: http://www.biolinguistics.eu/index.php/biolinguistics/article/view/415/361

    You point out that "The view Chomsky (and Berwick and Dawkins and Tatersall) favor is that there is something qualitatively different between language capable brains and ones that are not." What exactly it is about our brains which is qualitatively different is addressed here: https://www.ucl.ac.uk/pals/research/linguistics/research/uclwpl/wpl/16papers/UCLWPL_2016_Murphy.pdf

    ReplyDelete
  4. Two interrelated points on evolution:

    You criticized the the famous aphorism about nothing in biology making sense except in the light of evolution. To a certain extent, of course, you're right--pretty much any absolute statement is overbroad to at least some degree--but I think you missed the deeper meaning.

    When you're trying to understand something in the biological realm, an evolutionary perspective provides an indispensable and unifying framework. The heart and the eye are good examples. Take the human case: You can't get very far in exploring the mechanics of either organ before you start wondering why it works the way it does, because no sane entity would ever design it that way from scratch. The answer, of course, is that it *wasn't* designed from scratch; it was essentially built out of components already on hand. If you look at other hearts or eyes--taking a comparative-biological approach--you generally start to see how everything fits together and how you ended up with what you have in front of you. But that's not all: You can also use the comparative approach to generate hypotheses about what this or that part of the structure does. Now, at times you can usefully focus on the structure of some particular kind of receptor, but big-picture--which is what we're talking about here--you're going to try to fit it into some kind of evolutionary framework. Without that, you have knowledge but not much in the way of understanding, to use Feynman's distinction.

    Similarly, there is little use in trying to suss out the evolution of a particular trait in isolation. I'm not sure that you *can* do that, but I'm not sure it makes sense to try. Take Chomsky's claim that Merge (or whatever) arose from a single mutation. To me that's always seemed pretty vacuous; it doesn't really tell us anything. So, for instance, if you mutated the same gene in C. elegans or Pan troglodytes, you're not going to end up with a bunch of worms or chimps that can do Merge; they don't have the pre-existing framework for it. What gives you Merge (assuming there is such a thing) is that mutation plus a whole lot of other stuff that came before it. Oh, and that mutation also probably gave you a bunch of other stuff besides Merge, but that's a side point.

    What Chomsky is fumbling towards is what pretty much every undergrad fumbles towards when learning about evolution--the desire to find "the gene for X" or "the gene for Y." It takes a lot of time and effort to explain that that's only rarely a good way to think about things at the macro level.

    ReplyDelete
    Replies
    1. This comment has been removed by the author.

      Delete
    2. I don't deny that providing an evo perspective can be useful. But I would argue that providing an engineering analysis can be useful as well and that neither trumps the other. The heart is a pump, the arm a lever. This puts constraints on what kind of structure it can have just as much as the observation that it derives from this and that bits of prior structure. The idea that nothing can be understood without an evo account is not only over the top, it is largely false. We understood a lot about how the heart pumps blood in the system before we have any idea of its evo history. Indeed, Marr's program is built on the assumption that understanding the computational problem the system "solves" can tell us a lot about how it is structured even in the absence of any sense of how it historically came into being or the components used to construct it. So, I think that I get the "deeper" meaning I just don't agree with the line you advance.

      I would go further: unless you know what something does (the computational problem) then it will be hard to understand what the components are doing and why some things were recruited and others wouldn't be useful. So we can have the dictum: there is no biological understanding without an understanding of the computational/engineering problem being solved. Does this get you all the way? Nope, you need to knows something about the components at hand. But it gets you a long way. IMO the Dobzhansky quote is way over played, and a good thing too as we know very little about the evo history of most of what we are biologically interested in so if understanding and insight awaited an evo account, we would be waiting a really long time.

      I should note btw, that Darwin did not think that nothing in bio could be understood without evolution. That's why he looked for those examples where tinkering was of importance to make his case, where solutions were not "perfect." He worried about the latter a lot precisely because the argument from design was such a threat to his project. But this recognizes that quite often an argument from good design is what explains what we find, as Darwin well understood.

      More to come:

      Delete
    3. Here's the more to come:

      Second, I disagree about the "trait in isolation" claim. Yes you need to know how things interact. But it is worthwhile asking about a particular trait's evolution and to ask about it you need to specify this trait's structural properties (describe what it is). Then you can ask how IT arose, not otherwise. Absent a specification of the trait of interest, it seems to me that there is no evo question to ask. And there are specifications of traits that care about what other traits they are related to and some that care about the fine structure of the trait itself (e.g. arm as a lever vs as connected to shoulder).

      Last point: "What gives you Merge (assuming there is such a thing) is that mutation plus a whole lot of other stuff that came before it." No what gave you FL is Merge plus a lot of other stuff. It is an open question whether there is anything that precedes Merge, whether it has any biological pre-conditions or precursors. If it is a novelty then there is no precursor. Of course if Merge arose in worms or chimps then unless there is a whole lot more it would not lead to an FL. But it might well lead to a certain kind of recursive capacity possibly detectable depending to what it hooks up with. Merge is not the entirety of FL. It is a component part if Chomsky is right. It is responsible for allowing unbounded hierarchy of the type visible in language. It is possible that such hierarchies are generable and lead to different cognitive objects/capacities if hooked up to other "interfaces."

      So the question is whether the capacity for hierarchical recursion traces to a simple operation or is an interaction of several more primitive operations. It is conceivable that it is simple. If so, it won't have an "interesting" evo history. My friends in evo biology don't seem fazed by this possibility. It really depends on how "simple" Merge is. Of course, to repeat, Merge is not language. So even given Merge there is no guarantee that out pops something like human linguistic facility. But nobody thought otherwise, certainly not Chomsky, however much he might "fumble."

      Delete