David Berlinsky is editing a newish online magazine Inference that articles in which I have
mentioned in several previous posts. The latest issue is full of fun for
linguists as there are four articles of immediate relevance. Here’s the link for the issue. Let me
say a word or two about the pieces.
The first is an essay by Chomsky that goes over familiar
ground regarding the distinctive nature of human linguistic capacity. He
observes that this observation has a Cartesian pedigree and that language was
recognized as distinctive (all and only humans have it) and wondrous (it was
free and capable of expressing unboundedly many thoughts) and demanding of some
kind of explanation (it really didn’t fit in well with what was understood to
be the causal structure of the physical world) as early as it was noticed.
As Chomsky notes, Cartesians had relatively little of
substance to say about the underpinnings of this wondrous capacity, mainly because
the 17th century lacked the mathematical tools for the project. They
had no way of describing how it was possible to “make infinite use of finite
means” as von Humboldt put it (2). This changed in the 20th century
with Church, Godel, Post and Turing laying the foundations of computation
theory. This work “demonstrated how a finite object like the brain could
generate an infinite variety of expressions.” And as a result, “[i]t became
possible, for the first time, to address part of” the problem that the
Cartesians identified (2).
Note the ‘part of’ hedge. As Chomsky emphasizes, the problem
the Cartesians identified has two parts. The first, and for them the most
important feature, is the distinction between “inclined” vs “impelled” behavior
(3). Machines are impelled to act, never “inclined.” Humans, being free agents, are most often
“inclined” (though they can be “compelled” as well). Use of language is the
poster child for inclined behavior. Cartesians had no good understanding of the
mechanics of inclination. As Chomsky observes, more than 300 years later,
neither do we. As he puts it, language’s “free creative use remains a mystery,”
as does free action in general (e.g. raising one’s hand) (3).
The second part, one that computation theory has given us a
modest handle on, is the unbounded nature of the thoughts we can express. This
feature very much impressed Galileo and Arnauld & Lancelot and von Humboldt,
and it should impress you too! The “infinite variety” of meaningfully distinct
expressions characteristic of human language “surpasse[s] all stupendous
inventions” (1). Chomsky has redubbed
this feature of language “the Basic Property” (BP). BP refers to a property of
the human brain, “the language faculty,” and its capacity to “construct a
digitally infinite array of structured expressions” each of which “is
semantically interpreted as expressing a thought, and each can be externalized
by some sensory modality such as speech” (2).
BP is what GG has been investigating for the last 60 years or so. Quite
a lot has been discovered about it (and yes, there is still lots that we don’t
know!).
Chomsky emphasizes something that is worth reemphasizing:
these facts about language are not
news. That humans have linguistic
creativity in the two senses above should not really be a matter of dispute. That humans do language like no other
animal does should also be uncontroversial. How
we do this is a very tough question, only a small part (very small part) of
which we have managed to illuminate. It is sad that much debate still
circulates around the whether
question rather than the how. It is
wasted time.
An important theme in Chomsky’s essay turns on how the world
looks when we have no idea what’s up. Here is a quote that I believe all good
scientifically inclined GGers should have tattooed to themselves (preferably in
some discrete place) (3):
When understanding is thin, we
expect to see extreme variety and complexity.
Absolutely! Variety and complexity are hallmarks of
ignorance. And this is why progress and simplicity go hand in hand. And this is
why I have clasped to my heart Dresher’s apposite dictum: There should be only
two kinds of papers in linguistics: (i) papers that show that two things that
look completely different are roughly the same and (ii) papers that show that
two things that are roughly the same are in fact identical. These are the
papers that highlight our progressively deeper understanding. Complication is
often necessary, but it is progressive just in case it paves the way for
greater simplicity.
The unification and simplicity is, thus, a leading indicator
of scientific insight. Within linguistics it has a second function. It allows
one to start addressing the issue of how FL might have evolved. Here’s Chomsky:
In the analysis of the Basic
Property, we are bound to seek the simplest computational procedure consistent
with the data of language. Simplicity is implicit in the basic goals of
scientific inquiry. It has long been recognized that only simple theories can
attain a rich explanatory depth. “Nature never doth that by many things, which
may be done by a few,” Galileo remarked, and this maxim has guided the sciences
since their modern origins. It is
the task of the scientist to demonstrate this, from the motion of the planets,
to an eagle’s flight, to the inner workings of a cell, to the growth of
language in the mind of a child. Linguistics seeks the simplest theory for an
additional reason: it must face the problem of evolvability. Not a great deal
is known about the evolution of modern humans. The few facts that are well
established, and others that have recently been coming to light, are rather
suggestive. They conform to the conclusion that the language faculty is very simple;
it may, perhaps, even be computationally optimal, precisely what is suggested
on methodological grounds.
Unless FL is simpler than we have considered it to be up
till now (e.g. far simpler than say GBish models make it out to be) then there
is little chance that we will be able to explain its etiology. So there are
both general methodological grounds for wanting simple theories of FL and linguistic internal reasons for
hoping that much of the apparent complexity of FL is just apparent.
Chomsky’s piece proceeds by rehearsing in short form the
basic minimalist trope concerning evolvability. First, that we know little
about it and that we will likely not know very much about it ever. Second, that
FL is a true species property as the Cartesians surmised. Third, that FL has
not evolved much since humans separated. Fourth, that FL is a pretty recent
biological innovation. The third and fourth points are taken to imply that the
Basic Property aspect of FL must be pretty simple in the sense that what we see
today pretty well reflects the original evo innovation and so its properties are
physically simple in that they have not been shaped by the forces of selection.
In other words, what we see in BP is pretty much undistorted by the shaping
effects of evolution and so largely reflect the physical constraints that
allowed it to emerge.
All of this is by now pretty standard stuff, but Chomsky
tells it well here. He goes on to do what any such story requires. He tries to
illustrate how a simple system of the kind he envisions will have those
features that GG has discovered to be characteristic of FL (e.g. structure
dependence, unboundedly many discrete structures capable of supporting semantic
interpretation etc.). This second step is what makes MP really interesting. We
have a pretty good idea what kinds of things FL concerns itself with. That’s
what 60 years of GG research has provided. MP’s goal is to show how to derive
these properties from simpler starting points, the simpler the better. The
target of explanation (the explanadum) are the “laws” of GB. MP theories are
interesting to the degree that they can derive these “laws” from simpler more
principled starting points. And, that, Chomsky argues, is what what makes Merge
based accounts interesting, they derive features that we have every reason to
believe characterize FL.[1]
Two other papers in the issue address these minimalist
themes. The first is a review of the recent Berwick & Chomsky (B&C) book
Why only us. The second is a review
of a book on the origins of symbolic artifacts. Cederic Boeckx (CB) reviews
B&C. Ian Tatersall (IT) reviews the second. The reviews are in interesting
conflict.
The Boeckx review is quite negative, the heart of the criticism
being that asking ‘why only humans have language’ is the wrong question. What
makes it wrong? Well, frankly, I am not sure. But I think that the CB review
thinks that asking it endorses a form of “exceptional nativism” (7) that fails
to recognize “the mosaic character of language,” which, if I get the point,
implies eschewing “descent with modification” models of evolution (the gold
standard according to CB) in favor of “top-down, all-or-nothing” perspectives
that reject comparative cognition models (or any animal models), dismiss
cultural transmission as playing any role in explaining “linguistic complexity”
and generally take a jaundiced view of any evolutionary accounts of language
(7-8). I actually am skeptical regarding any of this.
Before addressing these points, however, it is interesting
that IT appears to take the position that CB finds wrong-headed. He thinks that
human symbolic capacities are biologically quite distinctive (indeed “unique”)
and very much in need of some explanation. Moreover, in contrast to CB, IT
thinks it pretty clear that this “symbolic activity” is of “rather recent
origin” and that, “as far as can be told, it was only our lineage that achieved
symbolic intelligence with all of its (unintended) consequences” (1). If we
read “symbolic” here to mean “linguistic” (which I think is a fair reading), it
appears that IT is asking for exactly the kind of inquiry that CB thinks
misconceived.
That said, let’s return to CB’s worries. The review makes
several worthwhile points. IMO, the two most useful are the observation that
there is more to language evolution than the emergence of the Basic Property
(i.e. Merge and discretely infinite hierarchically structured objects) and that
there may be more time available for selection to work its magic than is
presupposed. Let’s consider these points
in turn.
I think that many would be happy to agree that though BP is
a distinctive property of human language it may not be the only distinctive linguistic property. CB is right to observe that
if there are others (sometimes grouped together as FLW vs FLN) then these need
to be biologically fixed and that, to date, MP has had little to say about
these. One might go further; to date it is not clear that we have identified many properties of FLW at all.
Are there any?
One plausible candidate involves those faculties recruited
for externalization. It is reasonable to think that once FLN was fixed in the species,
that linking its products to the AP interface required some (possibly
extensive) distinctive biological retrofitting. Indeed, one might imagine that
all of phonology is such a biological kludge and that human phonology has no
close biological analogues outside of humans.[2]
If this is so, then the question of how much time this retrofitting required
and how fast the mechanisms of evolution (e.g. selection) operate is an
important one. Indeed, if there was special retrofitting for FLW linguistic
properties then these must have all taken place before the time that humans
went their separate ways for precisely the reasons that Chomsky likes to
(rightly) emphasize: not only can any human acquire the recursive properties of
any G, s/he can also acquire the FLW properties of any G (e.g. any phonology,
morphology, metrical system etc.).[3]
If acquiring any of these requires a special distinctive biology, then this
must have been fixed before we went our separate ways or we would expect,
contrary to apparent fact, that e.g. some “accents” would be inaccessible to
some kids. CB is quite right that it behooves us to start identifying
distinctive linguistic properties beyond the Basic Property and asking how they
might have become fixed. And CB is also right that this is a domain in which
comparative cognition/biology would be very useful (and has already been
started (see note 2). It is less clear that any of this applies to explaining
the evolution of the Basic Property itself.
If this is right, it is hard for me to understand CB’s
criticism of B&C’s identification of hierarchical recursion as a
very central distinctive feature of FL and asking how it could have
emerged. CB seems to accept this point
at times (“such a property unquestionably exists” (3)) but thinks that B&C
are too obsessed with it. But this seems to me an odd criticism. Why? Because B&C’s
way into the ling-evo issues is exactly the right
way to study the evolution of any
trait: First identify the trait of
interest. Second, explain how it could have emerged. B&C identify the trait (viz. hierarchical
recursion) and explain that it arose via the one time (non-gradual) emergence
of a recursive operation like Merge. The problem with lots of evo of lang work
is that it fails to take the first step of identifying the trait at issue. But
absent this any further evolutionary speculation is idle. If one concedes that a
basic feature of FL is the Basic Property, then obsessing about how it could have emerged is exactly the
right way to proceed.
Furthermore, and here I think that CB’s discussion is off
the mark, it seems pretty clear that this
property is not going to be all that amenable to any thing but a “top-down,
all-or-nothing” account. What I mean is that recursion is not something that
takes place in steps, a point that Dawkins made succinctly in support of
Chomsky’s proposal (see here).
As he notes, there is no such thing as “half recursion” and so there will be no
very interesting “descent with modification” account of this property. Something special happened in humans. Among other
things this led to hierarchical recursion. And this thing, whatever it was, likely
came in one fell swoop. This might not be all there is to say about language,
but this is one big thing about it and I don’t see why CB is resistant to this
point. Or, put another way, even if CB is right about many other features of
language being distinctive and amenable to more conventional evo analysis, it
does not gainsay the fact that the Basic Property is not one of these.
There is actually a more exorbitant possibility that perhaps
CB is reacting to. As the review notes (7): “Language is special, but not all
that special; all creatures have special abilities.” I don’t want to over-read
this, but one way of taking it is that different “abilities” supervene on
common capacities. This amounts to a warning not to confuse apparent
expressions of capacities for
fundamental differences in
capacities. This is a version of the standard continuity thesis (that
Lenneberg, among others, argued is very misleading (i.e. false) wrt language).
On this view, there is nothing much different in the capacities of the
“language ready” brain from the “language capable” brain. They are the same
thing. In effect, we need add nothing to an ape brain to get ours, though some
reorganization might be required (i.e no new circuits). I personally don’t
think this is so. Why? For the traditional reasons that Chomsky and IT note,
namely that nothing else looks like it does language like we do, even remotely.
And though I doubt that hierarchical recursion is the whole story (and have
even suggested that something other than Merge is the secret sauce that got
things going), I do think that it is a big part of it and that downplaying its
distinctiveness is not useful.
Let me put this another way. All can agree that evolution
involves descent with modification. The question is how big a role to attribute
to descent and how much to modification (as well as how much modification is
permitted). The MP idea can be seen as saying that much of FL is there before
Merge got added. Merge is the “modification” all else the “descent.” There will
fe features of FL continuous with what came before and some not continuous. No
mystery about the outline of such an analysis, though the details can be very
hard to develop. At any rate, it is hard for me to see what would go wrong if
one assumed that Merge (like the third color neuron involved in trichromatic
vision (thx Bill for this)) is a novel circuit and that FL does what it does by
combining the powers of this new operation with those cognitive/computational
powers inherited from our ancestors. That would be descent with modification.
And, so far as I can tell, that is what a standard MP story like that in
B&C aims to deliver. Why CB doesn’t like (or doesn’t appear to like) this
kind of story escapes me.
Observe that how one falls on the distinctiveness of BC
issue relates to what one thinks of the short time span observation (i.e.
language is of recent vintage so there is little time for natural selection or
descent with modification to work its magic). The view Chomsky (and Berwick and
Dawkins and Tatersall) favor is that there is something qualitatively different between language capable brains and ones
that are not. This does not mean that they don’t also greatly overlap. It just means that they are not capacity
congruent. But if there is a qualitative difference (e.g. a novel kind of
circuit) then the emphasis will be on the modifications, not the descent in
accounting for the distinctiveness. B&C is happy enough with the idea that
FL properties are largely shared with
our ancestors. But there is something different, and that difference is a big
deal. And we have a pretty good idea about (some of) the fine structure of that
difference and that is what
Minimalist linguistics should aim to explain.[4]
Indeed, I have argued and would continue to argue that the name of the
Minimalist game is to explain these very properties in a simple way. But I’ve
said that already here, so I won’t belabor the point (though I encourage you to
do so).
A few more random remarks and I am done. The IT piece
provides a quick overview of how distinctive human symbolic (linguistic?)
capacities are. In IT’s view, very. In IT’s view, the difference also emerged
very recently, and understanding that is critical to understanding modern
humans. And he is not alone. The reviewee Genevieve von Petziger appears to
take a similar view, dating the start of the modern human mind to about 80kya
(2). All this fits in with the dates that Chomsky generally assumes. It is nice
to see that (some) people expert in this area find these datings and the idea
that the capacity of interest is unique to us credible. Of course, to the
degree that this dating is credible and to the degree that this is not a long time for evolution to
exercise its powers the harder the evolutionary problem becomes. And, of
course, that’s what makes the problem interesting. At any rate, what the IT
review makes clear is that the way Chomsky has framed the problem is not
without reasonable expert support. Whether this view is correct, is, of course,
an empirical matter (and hence beyond my domain to competently judge).
Ok, let me mention two more intellectual confections of
interest and we are done. I will be short.
The first is a review of Wolfe’s book by David Lobina and
Mark Brenchley. It is really good and I cannot recommend it highly enough. I
urge you in particular to read the discussion on recursion as self-reference vs
self-embedding and the very enlightening discussion of how Post’s original
formalism (might have) led to some confusion on these issues. I particularly
liked the discussion of how Merge de-confuses them, in effect by dumping the
string based conception of recursion that Post’s formalism used (and which
invited a view of recursion as self-embedding) and implementing the recursive
idea more cleanly in a Merge like system in which linguistic structures are
directly embedded in one another without transiting through strings at all.
This cleanly distinguishes the (misleading) idea that the recursion lies with
embedding clauses within clauses from the more fundamental idea that recursion
requires some kind of inductive self-reference. Like I said, the discussion is
terrific and very useful.
And now for desert: read David Adger’s fun review of Arrival. I confess that I did not really
like the movie that much, but after reading David’s review, I intend to re-see
it with a more open mind.
That’s it. Take a look at the issue of Inference. It’s nice to see serious linguistic issues intelligently
discussed in a non-specialist’s venue. It can be done and done well. We need
more of it.
[1]
Chomsky also mentions that how lexical items have very distinctive properties
and that we understand very little about them. This ahs become a standard trope
in his essays, and a welcome one. It seems that lexical items are unlike animal
signs in that the latter are really “referential” in ways that the former are
not. The how and whys behind this, however, is completely opaque.
[3]
There is another possibility: once FLN is in place there is only one way to retrofit all the components
of FLW. If so, then there is no selection going on here and so the fact that
all those endowed with FLNs share common FLWs would not require a common ancestor
for the FLWs. Though I know nothing about these things, this option strikes me
as far-fetched. If it is, then the logic that Chomsky has deployed for arguing
that FLN was in place before humans went their separate ways would hold for FLW
as well.
[4]
CB makes a claim that is often mooted in discussions about biology. It is
Dobzhansky’s dictum that nothing in biology makes sense except in the light of
evolution. I think that this is overstated. Lots of biology “makes sense”
without worrying about origin. We can understand how hearts work or eyes see or
vocal tracts produce sounds without knowing anything at all about how they
emerged. This is not to diss the inquiry: we all want to know how things came
to be what they are. But the idea that natural selection is the only thing that
makes sense of what we see is often overblown, especially so when Dobzhansky
quotes are marshaled. For some interesting discussion of this see this.
>There should be only two kinds of papers in linguistics: (i) papers that show that two things that look completely different are roughly the same and (ii) papers that show that two things that are roughly the same are in fact identical.
ReplyDeleteSo all papers about E-languages are useless and should not exist? The entirety of the Labovian enterprise, say? Also the entirety of historical linguistics? Documentation and description of endangered languages is right out, I guess, given how it focus on surface, language-specific properties (a kind of data which, unlike the general faculty of language, will likely disappear within a generation)? No more papers on forensic linguistics? Natural language processing, data extraction from textual corpora, the cognition of reading and writing? Methods to help post-critical-period adults learn new languages better and faster? Everyone who studies these things should be defunded ASAP, their journals should go under etc.?
Or are we allowed to investigate these things, only we may not use the elevated term "linguistics" for them? "Linguistics" is reserved for the study of the biological faculty of language, and every other scientific investigation of language should be called some other, lesser thing?
You can investigate whatever you want. I am, sadly, not Tzar of what gets done in the field. My point was more modest: to the degree you are interested in the structure of FL (which I take to be the goal of work in GG) then these are the kinds of papers you are interested in as these are the kinds of papers that encourage explanation. Other work is interesting to the degree that it fosters this kind. So, yes, if you care to explain this is the kind of paper you will write. Of course, explanation isn't the only thing in life nor do I control funding/publication etc (nor would I want to despite my snide remark above). What I DO think is that for my concerns (which I admit are not widely shared) Dresher's point has hit a worthy nail right on the head.
DeleteJust a short point about L&B on recursion... It is s nice piece, and I think they are basically right, but there is a deep connection between recursion and self-embedding that they downplay. A recursive df. enumerates a set, where the members don't have to be specified as involving self-embedding. That's right, but such members understood as values of the function do involve embedding of the self-referred to concept qua enumerated. So, to use the arithmetic cases, instead of thinking of 1 and 2, say, as distinct objects, we think of them as s(0) and s(s(0)), respectively; mutates mutandis for the factorial, where 6, say, is (((1)x2)x3). Lord knows Wolfe knows shit, but the ideas are tricky.
ReplyDeleteGreat points--that was one of the problems I had with L&B's piece (which I didn't find all that helpful, to be honest). I think it would've been stronger if they'd drawn a clearer distinction between recursion and self-embedding, because their examples of recursion do involve embedding; of *course* factorial 4 has factorial 3, 2, and 1 embedded within. The decision to end on the Fitch quote seemed quite strange (I thought Fitch was there to look for recursion, not embedding).
DeleteThanks, Steven. Yes, I thought the distinction remains a bit unclear in the paper. Also, yes, not sure what the Fitch quote means. Is it that he was after evidence for embedding because that is what DE had erroneously conflated with recursion?
DeleteMy review of B&C raises a number of quite different objections, mainly focused on chapter 4: http://www.biolinguistics.eu/index.php/biolinguistics/article/view/415/361
ReplyDeleteYou point out that "The view Chomsky (and Berwick and Dawkins and Tatersall) favor is that there is something qualitatively different between language capable brains and ones that are not." What exactly it is about our brains which is qualitatively different is addressed here: https://www.ucl.ac.uk/pals/research/linguistics/research/uclwpl/wpl/16papers/UCLWPL_2016_Murphy.pdf
Two interrelated points on evolution:
ReplyDeleteYou criticized the the famous aphorism about nothing in biology making sense except in the light of evolution. To a certain extent, of course, you're right--pretty much any absolute statement is overbroad to at least some degree--but I think you missed the deeper meaning.
When you're trying to understand something in the biological realm, an evolutionary perspective provides an indispensable and unifying framework. The heart and the eye are good examples. Take the human case: You can't get very far in exploring the mechanics of either organ before you start wondering why it works the way it does, because no sane entity would ever design it that way from scratch. The answer, of course, is that it *wasn't* designed from scratch; it was essentially built out of components already on hand. If you look at other hearts or eyes--taking a comparative-biological approach--you generally start to see how everything fits together and how you ended up with what you have in front of you. But that's not all: You can also use the comparative approach to generate hypotheses about what this or that part of the structure does. Now, at times you can usefully focus on the structure of some particular kind of receptor, but big-picture--which is what we're talking about here--you're going to try to fit it into some kind of evolutionary framework. Without that, you have knowledge but not much in the way of understanding, to use Feynman's distinction.
Similarly, there is little use in trying to suss out the evolution of a particular trait in isolation. I'm not sure that you *can* do that, but I'm not sure it makes sense to try. Take Chomsky's claim that Merge (or whatever) arose from a single mutation. To me that's always seemed pretty vacuous; it doesn't really tell us anything. So, for instance, if you mutated the same gene in C. elegans or Pan troglodytes, you're not going to end up with a bunch of worms or chimps that can do Merge; they don't have the pre-existing framework for it. What gives you Merge (assuming there is such a thing) is that mutation plus a whole lot of other stuff that came before it. Oh, and that mutation also probably gave you a bunch of other stuff besides Merge, but that's a side point.
What Chomsky is fumbling towards is what pretty much every undergrad fumbles towards when learning about evolution--the desire to find "the gene for X" or "the gene for Y." It takes a lot of time and effort to explain that that's only rarely a good way to think about things at the macro level.
This comment has been removed by the author.
DeleteI don't deny that providing an evo perspective can be useful. But I would argue that providing an engineering analysis can be useful as well and that neither trumps the other. The heart is a pump, the arm a lever. This puts constraints on what kind of structure it can have just as much as the observation that it derives from this and that bits of prior structure. The idea that nothing can be understood without an evo account is not only over the top, it is largely false. We understood a lot about how the heart pumps blood in the system before we have any idea of its evo history. Indeed, Marr's program is built on the assumption that understanding the computational problem the system "solves" can tell us a lot about how it is structured even in the absence of any sense of how it historically came into being or the components used to construct it. So, I think that I get the "deeper" meaning I just don't agree with the line you advance.
DeleteI would go further: unless you know what something does (the computational problem) then it will be hard to understand what the components are doing and why some things were recruited and others wouldn't be useful. So we can have the dictum: there is no biological understanding without an understanding of the computational/engineering problem being solved. Does this get you all the way? Nope, you need to knows something about the components at hand. But it gets you a long way. IMO the Dobzhansky quote is way over played, and a good thing too as we know very little about the evo history of most of what we are biologically interested in so if understanding and insight awaited an evo account, we would be waiting a really long time.
I should note btw, that Darwin did not think that nothing in bio could be understood without evolution. That's why he looked for those examples where tinkering was of importance to make his case, where solutions were not "perfect." He worried about the latter a lot precisely because the argument from design was such a threat to his project. But this recognizes that quite often an argument from good design is what explains what we find, as Darwin well understood.
More to come:
Here's the more to come:
DeleteSecond, I disagree about the "trait in isolation" claim. Yes you need to know how things interact. But it is worthwhile asking about a particular trait's evolution and to ask about it you need to specify this trait's structural properties (describe what it is). Then you can ask how IT arose, not otherwise. Absent a specification of the trait of interest, it seems to me that there is no evo question to ask. And there are specifications of traits that care about what other traits they are related to and some that care about the fine structure of the trait itself (e.g. arm as a lever vs as connected to shoulder).
Last point: "What gives you Merge (assuming there is such a thing) is that mutation plus a whole lot of other stuff that came before it." No what gave you FL is Merge plus a lot of other stuff. It is an open question whether there is anything that precedes Merge, whether it has any biological pre-conditions or precursors. If it is a novelty then there is no precursor. Of course if Merge arose in worms or chimps then unless there is a whole lot more it would not lead to an FL. But it might well lead to a certain kind of recursive capacity possibly detectable depending to what it hooks up with. Merge is not the entirety of FL. It is a component part if Chomsky is right. It is responsible for allowing unbounded hierarchy of the type visible in language. It is possible that such hierarchies are generable and lead to different cognitive objects/capacities if hooked up to other "interfaces."
So the question is whether the capacity for hierarchical recursion traces to a simple operation or is an interaction of several more primitive operations. It is conceivable that it is simple. If so, it won't have an "interesting" evo history. My friends in evo biology don't seem fazed by this possibility. It really depends on how "simple" Merge is. Of course, to repeat, Merge is not language. So even given Merge there is no guarantee that out pops something like human linguistic facility. But nobody thought otherwise, certainly not Chomsky, however much he might "fumble."