The discussion of the SMT posts has gotten more abstract
than I hoped. The aim of the first post discussing the results by Pietroski,
Lidz, Halberda and Hunter was to bring the SMT down to earth a little and
concretize its interpretation in the context of particular linguistic investigations. PLHH investigate the following: there are
many ways to represent the meaning of most,
all of which are truth functionally equivalent. Given this, are the representations empirically equivalent or are
there grounds for arguing choosing one representation over the others. PLHH
propose to get a handle on this by investigating how these representations are used by the ANS+visual system in
evaluating dot scenes wrt statements like most
of the dots are blue. They discover that the ANS+visual system always uses
one of three possible representations to evaluate these scenes even when use of the others would be both
doable and very effective in that context. When one further queries the
core computational predilections of the ANS+visual system it turns out that the
predicates that it computes easily coincide with those that the “correct”
representation makes available. The conclusion is that the one of the three
representations is actually superior to the others qua linguistic representation of the meaning of most, i.e. it is the linguistic meaning of most. This all fits rather well with the SMT. Why?
Because the SMT postulates that one way of empirically evaluating candidate
representations is with regard to their fit
with the interfaces (ANS+visual) that use it. In other words, the SMT bids us
look to how grammars fit with interfaces and, as PLHH show, if one understands
‘fit’ to mean ‘be transparent with’ then one meaning trumps the others when we
consider how the candidates interact with the ANS+visual system.
It is important to note that things need not have turned out
this way empirically. It could have been the case that despite core capacities
of the ANS+visual system the evaluation procedure the interface used when evaluating
most sentences was highly context
dependent, i.e. in some cases it used the one-to-one strategy, in others the
‘|dots ∩ blue| - |dots ∩ not-blue|’ strategy and sometimes the ‘|dots ∩ blue| - [|dots| - |dots ∩ blue|]’ strategy. But, and this is important,
this did not happen. In all cases the
interface exclusively used the third option, the one that fit very snugly with
the basic operations of the ANS+visual system. In other words, the
representation used is the one that the SMT (interpreted as the Interface
Transparency Thesis) implicates. Score one for the SMT.
Note that the argument puts together various strands: it
relies on specific knowledge on how the ANS+visual system functions. It relies
on specific proposals for the meaning of most
and given these it investigates what happens when we put them together. The kicker is that if we assume that the
relation between the linguistic representation and what the ANS+visual system
uses to evaluate dot scenes is “transparent” then we are able to predict[1]
which of the three candidate representations will in fact be used in a
linguistic+ANS+visual task (i.e. the task of evaluating a dot scene for a given
most sentence[2]).[3]
The upshot: we are able to use information from how the
interface behaves to determine a property of a linguistic representation. Read that again slowly: PLHH argue that
understanding how these tasks are accomplished provides evidence for what the
linguistic meanings are (viz. what the correct representations of the meanings
are). In other words, experiments like this bear on the nature of linguistic
representations and a crucial assumption in tying the whole beautiful package
together is the SMT interpreted along the lines of the ITT.
As I mentioned in the first post on the SMT and Minimalism
(here), this is not the only exemplar of the SMT/ITT in action. Consider one
more, this time concentrating on work by Colin Phillips (here). As previously
noted (here), there are methods for tracking the online activities of parsers.
So, for example, the Filled Gap Effect (FGE) tracks the time course of mapping
a string of words into structured representations. Question: what rules do parsers use in doing
this. The SMT/ITT answer is that parsers use the “competence” grammars that
linguists with their methods investigate. Colin tests this by considering a very complex instance: gaps within
complex subjects. Let’s review the argument.
First some background.
Crain and Fodor (1985) and Stowe (1986) discovered that the online
process of relating a “filler” to its “gap” (e.g. in trying to assign a Wh a
theta role by linking it to its theta assigning predicate) is very eager. Parsers try to shove wayward Whs into
positions even if filled by another DP.
This eagerness shows up behaviorally as slowdowns in reading times when
the parser discovers a DP already homesteading in the thematic position it
wants to shove the un-theta marked DP into. Thus in (1a) (in contrast to (1b),
there is a clear and measurable slowdown in reading times at Bill because it is a place that the who could have received a theta role.
(1) a.
Who did you tell Bill about
b.
Who did you tell about Bill
Thus, given the parser’s eagerness, the FGE becomes a probe
for detecting linguistic structure built online. A natural question is where do
FGEs appear? In other words, do they “respect” conditions that “competence”
grammars code? BTW, all I mean by
‘competence grammars’ are those things that linguists have proposed using their
typical methods (one’s that some Platonists seem to consider the only valid
windows into grammatical structure!)? The
answer appears to be they do. Colin reviews the literature and I refer you to
his discussion.[4] How do FGEs show that parsers respect
grammatical structure? Well, they seem not
to apply within islands! In other words, parsers do not attempt to related Whs to gaps within islands. Why? Well given
the SMT/ITT it is because Whs could not have moved from positions wihin islands
and so they are not potential theta
marking sites for the Whs that the parser is eagerly trying to theta mark. In
other words, given the SMT/ITT we expect parser eagerness (viz. the FGE) to be
sensitive to the structure of grammatical representations, and it seems that it
is.
Observe again, that this is not a logical necessity. There
is no a priori reason why the grammars
that parsers use should have the properties that linguists have postulated,
unless one adopts the SMT/ITT that is. But let’s go on discussing Colin’s paper
for it gets a whole lot more subtle than this. It’s not just gross properties
of grammars that parsers are sensitive to, as we shall presently see.
Colin consider gaps within two kinds of complex subjects.
Both prevent direct extraction of a Wh (2a/3a), however, sentences like (2b)
license parasitic gaps while those like (3b) do not:
(2) a.
*What1 did the attempt to repair t1 ultimately damage the
car
b.
What1 did the attempt to repair t1 ultimately damage t1
(3) a.
*What1 did the reporter that criticized t1 eventually
praise the war
b. *What did the reporter that criticized
t1 eventually praise t1
So the grammar allows gaps related to extracted Whs in (2b)
but not (3b), but only if this is a parasitic gap. This is a very subtle set of grammatical
facts. What is amazing (in my view
nothing short of unbelievable) is that the parser respects these parasitic gap licensing conditions. Thus, what Colin shows is that we find FGEs at
the italicized expressions in (4a) but not (4b):
(3) a.
What1 did the attempt to repair the
car ultimately …
b.
What1 did the reporter that
criticized the war eventually …
This is a case where the parser is really tightly cleaving to distinctions that the grammar makes. It
seems that the parser codes for the possibility of a parasitic gap while
processing the sentence in real time.
Again, this argues for a very transparent relation between the
“competence” grammar and the parsing grammar, just as the SMT/ITT would
require.
I urge the interested to read Colin’s article in full. What
I want to stress here is that this is another concrete illustration of the
SMT. If
grammatical representations are optimal realizations of interface conditions
then the parser should respect the distinctions that grammatical
representations make. Colin presents evidence that it does, and does so very
subtly. If linguistic representations are used
by interfaces, then we expect to find this kind of correlation. Again, it is
not clear to me why this should be true given certain widely bruited Platonic
conceptions. Unless it is precisely these
representations that are used by the parser, why should the parser respect its
dicta? There is no problem understanding
how this could be true given a standard mentalist conception of grammars. And
given the SMT/ITT we expect it to be true. That we find evidence in its favor
strengthens this package of assumptions.
There are other possible illustrations of the SMT/ITT. We should develop a sense of delight at
finding these kind of data. As Colin’s stuff shows, the data is very complex
and, in my view, quite surprising, just like PLHH’s stuff. In addition, they
can act as concrete illustrations of how to understand the SMT in terms of
Interface Transparency. An added bonus
is that they stand as a challenge to certain kinds of Platonist conceptions, I
believe. Bluntly: either these
representations are cognitively available or we cannot explain why the
ANS+visual system and the parser act as if they were. If Platonic
representations are cognitively (and neurally, see note 3) available, then they
are not different from what mentalists have taken to be the objects of study
all along. If from a Platonist perspective they are not cognitively (and
neurally) available then Platonists and mentalists are studying different
things and, if so, they are engaged in parallel rather than competing
investigations. In either case, mentalists need take heed of Platonist results
exactly to the degree that they can be reinterpreted mentalistically.
Fortunately, many (all?) of their results can be so interpreted. However, where this is not possible, they would be of absolutely no interest to the project of describing linguistic competence. Just metaphysical curiosities
for the ontologically besotted.
[1]
Recall, as discussed here, ‘predict’ does not mean ‘explain.’
[2]
Remember, absent the sentence and in
specialized circumstances the visual system has no problem using strategies
that call on powers underlying the other two non-exploited strategies. It’s
only when the visual system is combined with the ANS and with the linguistic most sentence probe that we get the
observed results.
[3]
Actually, I overstate things here: we are able to predict some of the properties of the right representation, e.g. that it
doesn’t exploit negatively specified predicates or disjunctions of predicates.
[4]
Actually, there are several kinds of studies reviewed, only some of which
involve FGEs. Colin also notes EEG studies that show P600 effects when one has
a theta-undischarged Wh and one crosses into an island. I won’t make a big deal
out of this, but there is not exactly a dearth of neuro evidence available for
tracking grammatical distinctions. They
are all over the place. What we don’t have are good accounts of how brains implement grammars. We have
tons of evidence that brain responses track grammatical distinctions, i.e. that
brains respond to grammatical structures. This is not very surprising if you
are not a dualist. After all we have endless amounts of behavioral evidence
(viz. acceptability judgments, FGEs, eye movement studies, etc.) and on the
assumption that human behavior supervenes on brain properties it would be
surprising if brains did not distinguish what human subjects distinguish
behaviorally. I mention this only to state the obvious: some kinds of Platonism
should find these kinds of correlations challenging. Why should brains track
grammatical structure if these live in Platonic heavens rather than
brains? Just asking.