Wednesday, February 25, 2015

More reading for the curious

Deep learning: Here are two papers that investigate the “psychological reality” of  some popular deep learning models. They are particularly important for those wishing to borrow insights from this literature for cognitive ends. What the papers show is that it is possible to construct stimuli that the systems systematically classify (i.e. classify with very high confidence) as objects that no human would mistake them for.  Thus, deep neural networks are easily fooled: see http://arxiv.org/abs/1412.1897

These papers do something that linguists commonly do. The papers are about negative data. Negative data describe what humans do not do (e.g. native English speakers do not accept sentences like “*who did you meat a man who saw”). If deep learning models are to be understood as psychological theories, they need to agree both on the good and the bad data (i.e. on what we accept and reject). So far, much of the discussion has been on the positive capacities of such systems. They can be trained to spot a dog in a picture. However, these papers observe that current systems spot dogs that are not there, or, more precisely, categorize some picture as a dog photo that no human would so categorize. Or as the first paper puts it:

A recent study revealed that changing an image (e.g. a lion) in a way imperceptible to humans can cause a DNN [deep neural network NH] to label; the image as something else entirely (e.g. mislabeling a lion a library). Here we show a related result: it is easy to produce images that are completely unrecognizable to humans, but the state of the art DNNs believe to be recognizable objects with 99.99% confidence (e.g. labeling with certainty that white noise state is a lion).

Why do DNNs do this? Right now, it seems that nobody knows. Need I say that these are important results for the psychological “reality” of DNNs? As every linguist knows, explaining negative data is critical in evaluating any proposal aimed at describing our mental powers.

A note on evolution:
http://www.newyorker.com/news/daily-comment/evolution-catechism. This is an interesting discussion of the obvious political import that theories of evolution have in the US. The points are mainly obvious and congenial (to me). However, there is one distinction I would have made that Gopnik does not; the difference between the fact of evolution versus the centrality of the mechanism of natural selection (NS) as the prime causal force behind evolution. The fact is completely uncontroversial. Indeed, it was considered commonplace before Darwin, though Darwin did a lot to cement the truth of this fact. What is somewhat controversial today is how large a part NS plays in explaining this fact. All agree that it plays some role, the question is how big.

In many ways the recent Evo-Devo discoveries replay discussions similar to those in the early cog revolution. The Evo-Devo stuff suggests that the range of options that NS has to pick from is quite a bit narrower than earlier believed (viz. there are very few ways of doing anything (e.g. building an eye) and these tend to be strongly conserved over evolutionary time scales. Of course, the fewer the options available, the less one looks to NS to explain the outcome. Why? Because NS relies on the idea that were you are is heavily dependent on the path you took to getting there. But if the number of paths to get anywhere is very small in number then why you got to where you are is less dependent on a long series of linked choices than on the one or two you made at the very start. So, it is not that NS plays no role. Rather the importance of NS’s role depends on how wide the range of possibilities. The question is then not either/or but how much. And these are theoretical/empirical questions.

So, Gopnik is quite right that denying the fact of evolution is a nutty thing to do, sort of like denying that the earth is round or the earth orbits the sun. However, questioning the size of NS effects to evolutionary trajectories is not. NS is a theory. Evolution is the fact. As Gopnik notes, theories evolve and change. One of the changes being currently contemplated is that NS is a less potent factor than heretofore believed. Even a Republican can believe this in scientific good faith.

A neat paper on sources of evolution making a splash:
Some scene setting from Bob Berwick:
This is an interesting evolutionary finding because it uncovered a new mechanism by which evolution can work very quickly in a very complex setting.  We don’t know much about how complex organ systems evolve – the “major transitions in evolution” – like our brains. There are so many genes involved – how is it all put together without it getting all tangled up?  But now somebody’s got their foot in the door about one of the biggest transitions of all – how placental mammals evolved pregnancy, and went from laying eggs externally to growing them inside their bodies. Turns out that again (surprise!) this involved a set of regulatory genes, plus – the real surprise –what are called ‘transposons’ – bits of genes that can leap whole genes at a single bound and insert themselves even across chromosomes.[1] (Lynch et al., Ancient Transposable Elements Transformed the Uterine Regulatory Landscape and Transcriptome during the Evolution of Mammalian Pregnancy, Cell Reports (2015.[2] Apparently, the transposons donated regulatory elements to the genes that were recruited to alter the immune system so that the mother wouldn’t reject fetal embryos as foreign (remember a fetus has all those unknown genes from daddy). Lynch et al. demonstrated that this involved thousands of genes in a carefully coordinated orchestration led in part by the transposons, enabling exceptionally rapid evolution.  As Lynch notes, nobody expected to find that evolution could work this way to evolve large complex organ systems. Seems we still have a lot to learn about the basic evolutionary machinery, more than 150 years after Darwin.
More on Minds and Bodies: Those that enjoyed Chomsky’s discussion of the mind-body problem (here) (or should I say the non-existence of the problem given Newton’s excision of body from the equation) and were (rightly) dissatisfied with my discussion (here) might enjoy a real philosophical exposition of the state of the art by John Collins (here). It engages with lots of the philo literature on these matters and is eminently readable, even for linguists. He discusses and defends a position that he calls it “methodological naturalism,” which, if understood and adopted as the standard in the cog-neuro sciences (including linguistics) will remove most of the metaphysical and epistemological underbrush that hinders fruitful collaboration between linguistics and neuro-types. So, take a look and pass it onto your friends (and enemies) in the neurosciences who ignore most of what you have to say.




[1] First discovered by Barbara McClintock in corn, in the 1940s.  Nobody believed her at first, because it violated everything people thought they new about Mendelism and genes, but she hammered away at it.  Forty years later she got a Nobel prize.

4 comments:

  1. I don't know if the deep learning papers imply that deep neural networks as a class of models could never mimic human performance - I think what the papers show is that the particular architecture they tested has this issue. I'm not sure that there's any function that can't be computed by a "deep neural network", but I could be missing something. At any rate these papers are very useful - this seems like the kind of work that needs to be done if we eventually want to make those models more psychologically plausible.

    And a question: are there are any existing models that that achieve more human-like performance, or that don't have blind spots of this sort? Unfortunately I know next to nothing about computer vision and couldn't find relevant references in the papers. There's a (plausible) speculation at the end of the Nguyen et al paper that generative models may not suffer from this problem, but I don't see why generative models could not be implemented as "deep neural networks".

    ReplyDelete
    Replies
    1. I did not intend to imply that DL models could not be fixed. Rather, I was applauding the concern for negative data that these papers were investigating. If these models are to be interpreted psychologically, this kind of data seems very pertinent. As for DL networks as a class, I am in no position to know. They authors did note that these were two different very standard DL systems and that both seemed fool able in more or less the same way. They also noted possible ways of getting around this, maybe. This is all good. That's what they should be doing. Good for them.

      As to you second question, I doubt it but don't know. As I noted, these are the state of the art systems that they are playing with. But I am no expert either.

      Delete
  2. Did you mean this by John Collins? (there is no link in your text) https://www.academia.edu/11051565/Naturalism_without_Metaphysics

    ReplyDelete
    Replies
    1. I did indeed. Thx. I fixed the link. A long day.

      Delete