The discussion of the SMT posts has gotten more abstract
than I hoped. The aim of the first
post discussing the results by Pietroski,
Lidz, Halberda and Hunter was to bring the SMT down to earth a little and
concretize its interpretation in the context of particular linguistic investigations. PLHH investigate the following: there are
many ways to represent the meaning of
most,
all of which are truth functionally equivalent.
Given this, are the representations empirically equivalent or are
there grounds for arguing choosing one representation over the others. PLHH
propose to get a handle on this by investigating
how these representations are used by the ANS+visual system in
evaluating dot scenes wrt statements like
most
of the dots are blue. They discover that the ANS+visual system always uses
one of three possible representations to evaluate these scenes
even when use of the others would be both
doable and very effective in that context. When one further queries the
core computational predilections of the ANS+visual system it turns out that the
predicates that it computes easily coincide with those that the “correct”
representation makes available. The conclusion is that the one of the three
representations is actually superior to the others
qua linguistic representation of the meaning of
most, i.e. it
is the linguistic meaning of
most. This all fits rather well with the SMT. Why?
Because the SMT postulates that one way of empirically evaluating candidate
representations is with regard to their
fit
with the interfaces (ANS+visual) that use it. In other words, the SMT bids us
look to how grammars fit with interfaces and, as PLHH show, if one understands
‘fit’ to mean ‘be transparent with’ then one meaning trumps the others when we
consider how the candidates interact with the ANS+visual system.
It is important to note that things need not have turned out
this way empirically. It could have been the case that despite core capacities
of the ANS+visual system the evaluation procedure the interface used when evaluating
most sentences was highly context
dependent, i.e. in some cases it used the one-to-one strategy, in others the
‘|dots ∩ blue| - |dots ∩ not-blue|’ strategy and sometimes the ‘|dots ∩ blue| - [|dots| - |dots ∩ blue|]’ strategy. But, and this is important,
this did not happen. In all cases the
interface exclusively used the third option, the one that fit very snugly with
the basic operations of the ANS+visual system. In other words, the
representation used is the one that the SMT (interpreted as the Interface
Transparency Thesis) implicates. Score one for the SMT.
Note that the argument puts together various strands: it
relies on specific knowledge on how the ANS+visual system functions. It relies
on specific proposals for the meaning of
most
and given these it investigates what happens
when we put them together. The kicker is that if we assume that the
relation between the linguistic representation and what the ANS+visual system
uses to evaluate dot scenes is “transparent” then we are able to
predict
which of the three candidate representations will in fact be used in a
linguistic+ANS+visual task (i.e. the task of evaluating a dot scene for a given
most sentence
).
The upshot: we are able to use information from how the
interface behaves to determine a property of a linguistic representation. Read that again slowly: PLHH argue that
understanding how these tasks are accomplished provides evidence for what the
linguistic meanings are (viz. what the correct representations of the meanings
are). In other words, experiments like this bear on the nature of linguistic
representations and a crucial assumption in tying the whole beautiful package
together is the SMT interpreted along the lines of the ITT.
As I mentioned in the first post on the SMT and Minimalism
(here), this is not the only exemplar of the SMT/ITT in action. Consider one
more, this time concentrating on work by Colin Phillips (
here). As previously
noted (
here), there are methods for tracking the online activities of parsers.
So, for example, the Filled Gap Effect (FGE) tracks the time course of mapping
a string of words into structured representations. Question: what rules do parsers use in doing
this. The SMT/ITT answer is that parsers use the “competence” grammars that
linguists with their methods investigate. Colin tests this by considering a
very complex instance: gaps within
complex subjects. Let’s review the argument.
First some background.
Crain and Fodor (1985) and Stowe (1986) discovered that the online
process of relating a “filler” to its “gap” (e.g. in trying to assign a Wh a
theta role by linking it to its theta assigning predicate) is very eager. Parsers try to shove wayward Whs into
positions even if filled by another DP.
This eagerness shows up behaviorally as slowdowns in reading times when
the parser discovers a DP already homesteading in the thematic position it
wants to shove the un-theta marked DP into. Thus in (1a) (in contrast to (1b),
there is a clear and measurable slowdown in reading times at Bill because it is a place that the who could have received a theta role.
(1) a.
Who did you tell Bill about
b.
Who did you tell about Bill
Thus, given the parser’s eagerness, the FGE becomes a probe
for detecting linguistic structure built online. A natural question is where do
FGEs appear? In other words, do they “respect” conditions that “competence”
grammars code? BTW, all I mean by
‘competence grammars’ are those things that linguists have proposed using their
typical methods (one’s that some Platonists seem to consider the only valid
windows into grammatical structure!)? The
answer appears to be they do. Colin reviews the literature and I refer you to
his discussion.
How do FGEs show that parsers respect
grammatical structure? Well, they seem
not
to apply within islands! In other words, parsers do
not attempt to related Whs to gaps within islands. Why? Well given
the SMT/ITT it is because Whs could not have moved from positions wihin islands
and so they are
not potential theta
marking sites for the Whs that the parser is eagerly trying to theta mark. In
other words, given the SMT/ITT we expect parser eagerness (viz. the FGE) to be
sensitive to the structure of grammatical representations, and it seems that it
is.
Observe again, that this is not a logical necessity. There
is no a priori reason why the grammars
that parsers use should have the properties that linguists have postulated,
unless one adopts the SMT/ITT that is. But let’s go on discussing Colin’s paper
for it gets a whole lot more subtle than this. It’s not just gross properties
of grammars that parsers are sensitive to, as we shall presently see.
Colin consider gaps within two kinds of complex subjects.
Both prevent direct extraction of a Wh (2a/3a), however, sentences like (2b)
license parasitic gaps while those like (3b) do not:
(2) a.
*What1 did the attempt to repair t1 ultimately damage the
car
b.
What1 did the attempt to repair t1 ultimately damage t1
(3) a.
*What1 did the reporter that criticized t1 eventually
praise the war
b. *What did the reporter that criticized
t1 eventually praise t1
So the grammar allows gaps related to extracted Whs in (2b)
but not (3b), but only if this is a parasitic gap. This is a very subtle set of grammatical
facts. What is amazing (in my view
nothing short of unbelievable) is that the parser respects these parasitic gap licensing conditions. Thus, what Colin shows is that we find FGEs at
the italicized expressions in (4a) but not (4b):
(3) a.
What1 did the attempt to repair the
car ultimately …
b.
What1 did the reporter that
criticized the war eventually …
This is a case where the parser is really tightly cleaving to distinctions that the grammar makes. It
seems that the parser codes for the possibility of a parasitic gap while
processing the sentence in real time.
Again, this argues for a very transparent relation between the
“competence” grammar and the parsing grammar, just as the SMT/ITT would
require.
I urge the interested to read Colin’s article in full. What
I want to stress here is that this is another concrete illustration of the
SMT. If
grammatical representations are optimal realizations of interface conditions
then the parser should respect the distinctions that grammatical
representations make. Colin presents evidence that it does, and does so very
subtly. If linguistic representations are used
by interfaces, then we expect to find this kind of correlation. Again, it is
not clear to me why this should be true given certain widely bruited Platonic
conceptions. Unless it is precisely these
representations that are used by the parser, why should the parser respect its
dicta? There is no problem understanding
how this could be true given a standard mentalist conception of grammars. And
given the SMT/ITT we expect it to be true. That we find evidence in its favor
strengthens this package of assumptions.
There are other possible illustrations of the SMT/ITT. We should develop a sense of delight at
finding these kind of data. As Colin’s stuff shows, the data is very complex
and, in my view, quite surprising, just like PLHH’s stuff. In addition, they
can act as concrete illustrations of how to understand the SMT in terms of
Interface Transparency. An added bonus
is that they stand as a challenge to certain kinds of Platonist conceptions, I
believe. Bluntly: either these
representations are cognitively available or we cannot explain why the
ANS+visual system and the parser act as if they were. If Platonic
representations are cognitively (and neurally, see note 3) available, then they
are not different from what mentalists have taken to be the objects of study
all along. If from a Platonist perspective they are not cognitively (and
neurally) available then Platonists and mentalists are studying different
things and, if so, they are engaged in parallel rather than competing
investigations. In either case, mentalists need take heed of Platonist results
exactly to the degree that they can be reinterpreted mentalistically.
Fortunately, many (all?) of their results can be so interpreted. However, where this is not possible, they would be of absolutely no interest to the project of describing linguistic competence. Just metaphysical curiosities
for the ontologically besotted.
I believe this case can serve to illustrate some of Norbert's points (even though in this case, the professor in question WAS fired). My impression is that social psychology is even worse than economics in its equating science with finding correlations in large data sets: especially if you read Stapel's book it becomes clear that social psychology is all about doing experiments and interpreting them in the 'right' way statistically, and hardly ever about trying to construct a theory with some explanatory depth.
If Stapel's research would not have been fraudulent, not many things would have changed. He found correlations between eating meat and being aggressive, or between seeing the word 'capitalism' and eating M&M's. In this way, he became an academic superstar, at least at the Dutch scale: he published in Science, was a Dean at Tilburg University (where as you may know, a thriving department of linguistics has been closed a few years ago because of its unfeasibility) and appeared on tv a lot with the outcomes of his 'research'.
People are now discussing what should be the consequences of this. The die-hard empiricists say that experiments should be more standardly replicated, we should do statistics on data sets just to see how likely it is that they have been made up, etc. But it seems to me that having a good and solid theory, or a number of competing theories, also helps here.
The point is, once you have a theory, some data can be PROBLEMATIC (or 'funny-looking', as Norbert says) for somebody believing in that theory, so that person will become suspicious and therefore motivated to replicate the experiments, or at least check all the relevant data. This apparently is hardly ever the case in social psychology: the (fabricated) observation that people who see the word 'capitalism' eat more M&Ms was just not problematic for anybody, since nobody had any deep expectations about the relation between seeing that word and consuming chocolate sweets to begin with.
But to be fair, it has to be noted that in this case after a number of years a few junior researchers were brave enough to discover the fraud and talk to the rector about it, and the guy was fired. (A detail which might interest linguists, and which is not mentioned in the NYT article, is that the committee which examined the fraud was led by the well-known psycholinguist Willem Levelt.) And that might shed some light on the asymmetry between the Hauser case and the RR case. The differences might have less to do with issues of methodology than with prestige and political power.
(I have to admit that I know much more about the Stapel case than about Hauser or RR.)