The Dynamics of Iterated
Learning
|
|
|
Simon Kirby |
|
LEC, Edinburgh |
What is Iterated
Learning?
|
|
|
An approach to understanding language
evolution |
|
A way of relating properties of the
learners and users of language to language structure |
|
A type of cultural evolution |
|
|
|
The mechanism whereby knowledge is
passed-on through observation of behaviour resulting from that knowledge |
Iterated learning as
evolutionary system
|
|
|
There are important differences between
iterated learning and biological evolution |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
The main difference: possibility of
discontinuity |
|
The main similarity: link between
stability and reproductive success? |
|
Question: is there selection (for
learnability and/or usability)? |
Iterated learning models
|
|
|
|
Two broad classes: |
|
Complex simulation models – target
specific features of language |
|
Minimal idea models – understand
evolutionary principles |
|
Components of ILMs: |
|
Learning mechanism |
|
Production mechanism |
|
Invention mechanism |
|
What is being learned? |
|
Sets of strings |
|
Parameter settings |
|
Morphology |
|
Meaning-signal mappings |
Main ILM results so far
|
|
|
|
|
Languages can adapt to pressures on
production and parsing |
|
Generalisations are better replicators |
|
Given structured meaning-spaces: |
|
Compositional languages are more stable |
|
Compositional languages can emerge with
invention. Depends on: |
|
Size of signal space |
|
Size of “bottleneck” |
|
Population structure |
|
Environment structure |
|
Meaning-space structure |
What can we conclude from
ILM results?
|
|
|
|
General methodological approach: |
|
Discover which simulation parameters
lead to language-like systems |
|
Hypothesise that these parameters are
necessary (and sufficient?) precursors for language |
|
|
|
Sufficiency conditions: tells us how
much we need to explain |
|
Necessity conditions: helps solve the uniqueness
problem. |
|
|
Example: structured
meaning-spaces
|
|
|
|
How does the structure of meaning-space
affect the structure of the languages that emerge in an ILM? |
|
Henry’s approach: |
|
Construct languages that reflect
structure of meaning-spaces |
|
Compare the expressivity of these
languages after learning |
|
Equate expressivity and stability
through iterated learning |
|
Languages that reflect highly
structured meaning-spaces most stable |
|
General idea: highly-structured meaning
spaces are part of our biological endowment. |
Two problems…
|
|
|
|
Model looks at what happens to a
language that is completely compositional vs. completely holistic |
|
i.e., Lyapunov local stability |
|
Note: same problem for Nowak et. al in Nature |
|
What about partially compositional
systems? |
|
Model assumes that meaning-space is
fixed and provided by some other evolutionary mechanism |
|
Seems to run counter to the “spirit” of
adaptation through iterated learning |
|
What if the meaning-space coevolves
with the syntactic system? |
|
|
A really simple model
|
|
|
What is language? |
|
A system for mediating between
interfaces |
Meanings
|
|
|
Simple feature-vectors |
|
Multiple spaces |
|
Many-to-one mapping from meanings to
objects in environment |
Objects as sets of
feature-vectors
|
|
|
Each object is associated with one
feature-vector per meaning space. |
|
|
|
|
|
|
|
|
|
|
|
|
|
Key point: agents are prompted to
express signals for each object (not each meaning!) |
Big Assumption
|
|
|
|
If a learner is presented with a
language that is compositional for
some part of meaning space,
then the learner will learn a compositional system. |
|
|
|
Can be justified on MDL grounds? |
|
Model of learning simply needs to keep
track of each feature-value heard. |
|
Exactly the same as Henry’s model |
|
Stability relates to the probability of
hearing each feature-value |
|
|
Differences…
|
|
|
|
Potentially more than one subsystem
operating at the same time |
|
Only looking at the stability of
compositional systems |
|
BUT… holistic system is identical to a
1-dimensional compositional system |
|
|
|
What this allows us to do: |
|
Deal with mixed systems (i.e., move
beyond Lyapunov) |
|
Look at evolution of meaning-space
usage |
Simulation
|
|
|
|
Agent knowledge:
list of feature-values |
|
Production:
prompted by an object, pick any way of expressing it |
|
An object can be expressed in a meaning
space if the corresponding feature vector <v1, v2, …> contains only
feature-values that the agent knows how to express. |
|
Learning:
add the feature-values heard to knowledge |
|
|
|
Parameters:
environment, meaning spaces, bottleneck |
Experiments
|
|
|
|
|
Stability of different meaning-spaces |
|
Relative stability of different
meaning-spaces against holistic system |
|
Complete dynamics with two
meaning-spaces |
|
Iterated learning with multiple
meaning-spaces |
|
|
|
|
|
What next? |
Holistic vs.
Compositional
Holistic with differing
bottlenecks
Differing meaning-spaces
Relative stability for
differing meaning spaces (compared with holistic)
Iterated learning with
mixed system
Complete dynamics for
mixed system
Invention
Effect of low bottleneck
Low bottleneck with
invention
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Multiple systems
Inexpressive
meaning-spaces
Problem – no pressure for
expressivity
|
|
|
The key bias: avoidance of many to one
mapping between meanings and signals |
|
BUT – many to one mapping between
environment and meanings |
Solution – discrimination
and contexts
|
|
|
Currently: speakers only presented with
target object |
|
Add: a context of other objects |
|
Obverter: only produce signals that
will discriminate target from context |
|
|
Selection for contextual
discrimination
Summary
|
|
|
We can extend Henry’s model to
mixed-systems |
|
Linking objects to multiple
meaning-structures leads to competition between meaning-spaces |
|
Evolution through Iterated Learning
extends to semantics |
|
Key point: |
|
|
|
|
What’s next?
|
|
|
|
Environment is random and smooth |
|
Adaptation of semantics to environment
structure? |
|
|
|
Problem: why aren’t languages all
binary? |
|
|
|
Populations |
|
Need to add explicit lexicon |
|
|
|
Still need more realistic model of
meaning flexibility |
Final questions