Keyboard Shortcuts?

×
  • Next step
  • Previous step
  • Skip this slide
  • Previous slide
  • mShow slide thumbnails
  • nShow notes
  • hShow handout latex source
  • NShow talk notes latex source

Click here and press the right key for the next slide (or swipe left)

also ...

Press the left key to go backwards (or swipe right)

Press n to toggle whether notes are shown (or add '?notes' to the url before the #)

Press m or double tap to slide thumbnails (menu)

Press ? at any time to show the keyboard shortcuts

 

Core System vs Module

The problems with core knowledge look like they might be the sort of problem a philospoher might be able to help with.
Jerry Fodor has written a book called 'Modularity of Mind' about what he calls modules. And modules look a bit like core systems, as I'll explain. Further, Spelke herself has at one point made a connection. So let's have a look at the notion of modularity and see if that will help us.

core system = module?

‘In Fodor’s (1983) terms, visual tracking and preferential looking each may depend on modular mechanisms.’

\citep[p.\ 137]{spelke:1995_spatiotemporal}

Spelke et al 1995, p. 137

So what is a modular mechanism?
\subsection{Modularity}
Fodor’s three claims about modules:
\begin{enumerate}
\item they are ‘the psychological systems whose operations present the world to thought’;
\item they ‘constitute a natural kind’; and
\item there is ‘a cluster of properties that they have in common’ \citep[p.\ 101]{Fodor:1983dg}.
\end{enumerate}

Modules

  1. they are ‘the psychological systems whose operations present the world to thought’;
  2. they ‘constitute a natural kind’; and
  3. there is ‘a cluster of properties that they have in common … [they are] domain-specific computational systems characterized by informational encapsulation, high-speed, restricted access, neural specificity, and the rest’ (Fodor 1983: 101)
Modules are widely held to play a central role in explaining mental development and in accounts of the mind generally.
Jerry Fodor makes three claims about modules:
What are these properties?
Properties of modules:
\begin{itemize}
\item domain specificity (modules deal with ‘eccentric’ bodies of knowledge)
\item limited accessibility (representations in modules are not usually inferentially integrated with knowledge)
\item information encapsulation (roughly, modules are unaffected by general knowledge or representations in other modules)
\item innateness (roughly, the information and operations of a module not straightforwardly consequences of learning; but see \citet{Samuels:2004ho}).
\end{itemize}
  • domain specificity

    modules deal with ‘eccentric’ bodies of knowledge

  • limited accessibility

    representations in modules are not usually inferentially integrated with knowledge

  • information encapsulation

    roughly, modules are unaffected by general knowledge or representations in other modules

    For something to be informationally encapsulated is, roughly, for its operation to be unaffected by the mere existence of general knowledge or representations stored in other modules (Fodor 1998b: 127)
  • innateness

    roughly, the information and operations of a module not straightforwardly consequences of learning

Domain specificity
limited accessibility
Let me illustrate limited accessibility ...
Limited accessbility is a familar feature of many cognitive systems. When you grasp an object with a precision grip, it turns out that there is a very reliable pattern. At a certain point in moving towards it your fingers will reach a maximum grip aperture which is normally a certain amount wider than the object to be grasped, and then start to close. Now there's no physiological reason why grasping should work like this, rather than grip hand closing only once you contact the object. Maximum grip aperture shows anticipation of the object: the mechanism responsible for guiding your action does so by representing various things including some features of the object. But we ordinarily have no idea about this. The discovery of how grasping is controlled depended on high speed photography. This is an illustration of limited accessibility. (This can also illustrate information encapsulation and domain specificity.)

maximum grip aperture

(source: Jeannerod 2009, figure 10.1)

Glover (2002, figure 1a)

Illusion sometimes affects perceptual judgements but not actions: information is in the system; information is not available to knowledge \citep{glover:2002_visual}.
See further \citet{bruno:2009_when}: They argue that Glover & Dixon's model \citep{glover:2002_dynamic} is incorrect, at least for grasping (pointing is a different story), because it predicts that the presence or absence of visual information during grasping shouldn't matter. But it does.

van Wermeskerken et al 2013, figure 1

You also get evidence for information encapsulation in four month olds. To illustrate, consider \citet{vanwermeskerken:2013_getting} ...
A Ponzo-like background can make one frog further away than the other.

van Wermeskerken et al 2013, figure 2

This affects which object four-month olds reach for, but does not affect the kinematics of their reaching actions.
What are these properties?
  • domain specificity

    modules deal with ‘eccentric’ bodies of knowledge

  • limited accessibility

    representations in modules are not usually inferentially integrated with knowledge

  • information encapsulation

    roughly, modules are unaffected by general knowledge or representations in other modules

    For something to be informationally encapsulated is, roughly, for its operation to be unaffected by the mere existence of general knowledge or representations stored in other modules (Fodor 1998b: 127)
  • innateness

    roughly, the information and operations of a module not straightforwardly consequences of learning

Information encapsulation
Innateness
So these are the key properties associated with modularity.
We've seen something like this list of properties before ... Compare the notion of a core system with the notion of a module
The two definitions are different, but the differences are subtle enough that we don't want both. My recommendation: if you want a better definition of core system, adopt core system = module as a working assumption and then look to research on modularity because there's more of it.

‘core systems are

  1. largely innate,
  2. encapsulated, and
  3. unchanging,
  4. arising from phylogenetically old systems
  5. built upon the output of innate perceptual analyzers’

(Carey and Spelke 1996: 520)

Modules are ‘the psychological systems whose operations present the world to thought’; they ‘constitute a natural kind’; and there is ‘a cluster of properties that they have in common’

  1. innateness
  2. information encapsulation
  3. domain specificity
  4. limited accessibility
  5. ...

core system = module ?

I think it is reasonable to identify core systems with modules and to largely ignore what different people say in introducing these ideas. The theory is not strong enough to support lots of distinctions.

Will the notion of modularity help us in meeting the objections to the Core Knowledge View?

Recall that the challenges were these:
  • multiple definitions
  • justification for definition by list-of-features
  • definition by list-of-features rules out explanation
  • mismatch of definition to application
Let’s go back and see what Fodor says about modules again ...
Consider the first objection, that there are multiple defintions ...
Not all researchers agree about the properties of modules. That they are informationally encapsulated is denied by Dan Sperber and Deirdre Wilson (2002: 9), Simon Baron-Cohen (1995) and some evolutionary psychologists (Buller and Hardcastle 2000: 309), whereas Scholl and Leslie claim that information encapsulation is the essence of modularity and that any other properties modules have follow from this one (1999b: 133; this also seems to fit what David Marr had in mind, e.g. Marr 1982: 100-1). According to Max Coltheart, the key to modularity is not information encapsulation but domain specificity; he suggests Fodor should have defined a module simply as 'a cognitive system whose application is domain specific' (1999: 118). Peter Carruthers, on the other hand, denies that domain specificity is a feature of all modules (2006: 6). Fodor stipulated that modules are 'innately specified' (1983: 37, 119), and some theorists assume that modules, if they exist, must be innate in the sense of being implemented by neural regions whose structures are genetically specified (e.g. de Haan, Humphreys and Johnson 2002: 207; Tanaka and Gauthier 1997: 85); others hold that innateness is 'orthogonal' to modularity (Karmiloff-Smith 2006: 568). There is also debate over how to understand individual properties modules might have (e.g. Hirschfeld and Gelman 1994 on the meanings of domain specificity; Samuels 2004 on innateness).
In short, then, theorists invoke many different notions of modularity, each barely different from others. You might think this is just a terminological issue. I want to argue that there is a substantial problem: we currently lack any theoretically viable account of what modules are. The problem is not that 'module' is used to mean different things-after all, there might be different kinds of module. The problem is that none of its various meanings have been characterised rigorously enough. All of the theorists mentioned above except Fodor characterise notions of modularity by stipulating one or more properties their kind of module is supposed to have. This way of explicating notions of modularity fails to support principled ways of resolving controversy.
No key explanatory notion can be adequately characterised by listing properties because the explanatory power of any notion depends in part on there being something which unifies its properties and merely listing properties says nothing about why they cluster together.
So much the same objections which applied to the very notion of core knowledge appear to recur for module. But note one interesting detail ...

Modules

  1. they are ‘the psychological systems whose operations present the world to thought’;
  2. they ‘constitute a natural kind’; and
  3. there is ‘a cluster of properties that they have in common … [they are] domain-specific computational systems characterized by informational encapsulation, high-speed, restricted access, neural specificity, and the rest’ (Fodor 1983: 101)

Will the notion of modularity help us in meeting the objections to the Core Knowledge View?

  • multiple definitions
  • justification for definition by list-of-features
  • definition by list-of-features rules out explanation
  • mismatch of definition to application
We’ve been considering the first objection, that there are multiple defintions ...
What about the objection that picking out a set of features is unjustified ...

Modules

  1. they are ‘the psychological systems whose operations present the world to thought’;
  2. they ‘constitute a natural kind’; and
  3. there is ‘a cluster of properties that they have in common … [they are] domain-specific computational systems characterized by informational encapsulation, high-speed, restricted access, neural specificity, and the rest’ (Fodor 1983: 101)
Interestingly, Fodor doesn't define modules by specifying a cluster of properties (pace Sperber 2001: 51); he mentions the properties only as a way of gesturing towards the phenomenon (Fodor 1983: 37) and he also says that modules constitute a natural kind (see Fodor 1983: 101 quoted above).

Will the notion of modularity help us in meeting the objections to the Core Knowledge View?

  • multiple definitions
  • justification for definition by list-of-features
  • definition by list-of-features rules out explanation
  • mismatch of definition to application
We’ve been considering the first objection, that there are multiple defintions ...
Same point applies to the claim that defining module by listing features is unexplanatory: if we are not listing features but identifying a natural kind, then the objection doesn’t quite get strarted.
As far as the ‘justification for definition by list-of-features’ and ‘definition by list-of-features rules out explanation’ problems go, everything rests on the idea that modules are a natural kind. I think this idea deserves careful scruitiny but as far as I know there's only one paper on this topic, which is by me. I'm not going to talk about the paper here; let me leave it like this: if you want to invoke a notion of core knowledge or modularity, you have to reply to these problems. And one way to reply to them--- the only way I know---is to develop the idea that modules are a natural kind. If you want to know more ask me for my paper and I'll send it to you.
Recall the discrepancy in looking vs search measures. What property of modules could help us to explain it?

Spelke et al 1992, figure 2

Hood et al 2003, figure 1

occlusion endarkening
violation-of-expectations

Charles & Rivera (2009)

  • domain specificity

    modules deal with ‘eccentric’ bodies of knowledge

  • limited accessibility

    representations in modules are not usually inferentially integrated with knowledge

  • information encapsulation

    roughly, modules are unaffected by general knowledge or representations in other modules

    For something to be informationally encapsulated is, roughly, for its operation to be unaffected by the mere existence of general knowledge or representations stored in other modules (Fodor 1998b: 127)
  • innateness

    roughly, the information and operations of a module not straightforwardly consequences of learning

We already considered innateness and inforamtion encapsulation
To say that a system or module exhibits limited accessibility is to say that the representations in the system are not usually inferentially integrated with knowledge.
This is a key feature we need to assign to modular representations (=core knowledge) in order to explain the apparent discrepancies in the findings about when knowledge emerges in development.
Limited accessibility explains why the representations might drive some actions (e.g. certain looking behaviours) but not others (e.g. certain searching actions).
But't the bare appeal to limited accessibility leaves open why the looking and not the searching (rather than conversely). Further, we have to explain why not searching with occlusion whereas not looking with endarkening. Clearly we can’t explain this pattern just by invoking information encapsulation.
And, of course, to say that we can explain something by invoking information encapsulation is too much. After all, limited accessibility is more or less what we're trying to explain. But this is the first problem --- the problem with the standard way of characterising modularity and core systems merely by listing features.

core system = module

Some, not all, objections to the Core Knowledge View overcome:

  • multiple definitions
  • justification for definition by list-of-features
  • definition by list-of-features rules out explanation
  • mismatch of definition to application

The Core Knowledge View
generates
no
relevant predictions.

The main objection is unresolved