610 likes | 702 Views
The Use of Computer Simulation in Studying Biological Evolution: Pure Possible Processes, Selective Regimes and Open-ended Evolution. Philippe Huneman IHPST (CNRS / Université PARIS I Sorbonne). The interplay between evolutionary biology and computer science Bioinformatics , Biocomputation
E N D
The Use of Computer Simulation in Studying Biological Evolution: Pure PossibleProcesses, Selective Regimes and Open-ended Evolution Philippe HunemanIHPST (CNRS / Université PARIS I Sorbonne)
The interplaybetweenevolutionarybiology and computer science • Bioinformatics, Biocomputation • Geneticalgorithms (lexicon : « crossing over », genesetc) • The radical AL claim : digital organisms are themselves living entities (rather than their simulations) (Adami 2002). Idea of non-carbon-basedforms of life as the essence of life • Version # 2: Darwinian evolution is an « algorithm » (Dennett, see Maynard Smith 2000)
What are simulations doing / teaching ? • What is the role of natural selection in them ? • Investigates the relations between biological evolution and computer simulations of evolving entities through natural selection.
A. Kinds of role of selection: a1. formalselectioncontext Todd and Miller (1995) on sexualselection (sexual selection is more an incentive for exploration than natural selection, since females, through mate choice, internalize the constraints of natural selection. ) Maley on biodiversity (emergence of new species is more conditioned by geographical barriers than by adaptive potential)
A2. No selectioncontext Boids (Reynolds) Chu and Adami (2000): simulation of phylogenies whose parameter is the mean number of same-order daughter families of a family Mc Shea (1996, 2005) : increase of complexity with no selection
B.Use: weak and strong • B1. Weak : model is used to test one hypothesis on one process; it simulates the behavior of the entities (boids; Maley’s biodiversity; etc.). Way to test hypotheses about the world
B2 Strong : entities of the model don’t correspond to real entities; the simulation ismeant to explore the kinds of behaviors of digital entitiesthemselves (Ray’sTierra, Holland’s Echo etc.). Hypotheses are made about the model itself “digital organisms” as “defined by the sequence of instructions that constitute their genome, are not simulated: they are physically present in the computer and live there” (Adami, 2002)
Echo : unrealistic assumptions concerning reproduction, absence of isolative reproduction for species, makes Echo a poor model of evolutionary biology (Crubelier 1997)
In the (weak or strong) simulations : causal processes (i.e. counterfactual dependencies between classes of sets of cells, and global state at next step).
a1 a 2 ….. a1+P a2+P ……………… ak ….. A1 n, j A2 n, j……… Ak n, j ….. b1n+1 b2n+1 …………… bkn+1.. Property P n at Step n (Add (on i) Disj (on j) Ai n, j ) Property Pn+1 at Step n+1 (all the bin+1 ) “If P n had not been the case, Pn+1 would not have been the case.” Causation as counterfactual dependence between steps in Cas Huneman Minds and machines 2008
-> In « formal selection » contexts simulations : those causal processes are actual « selective processes »
Yet the entities in the simulations can not exactly match biological entities: In Echo, you don’t have species easily, in Tierra no lineages etc. If one system is designed to study some level of biological reality, the other levels are not ipso facto given (whereas if you have, e.g., organisms you have genes and species etc.)
-> In actual biology : all levels of the hierarchies are acting together • So computer simulations display « pure possible processes » concerning the entities modelised, located in a target-level of the hierarchy (no implicit entangling between levels)
In the case of formal selection simulations, « pure » selective processes occur • Ex. of « natural selection » sensu Channon : Echo or Hillis’s coevolution between sorting problems: « natural selection » simulations. Yet in Echo, for ex., the class of possible actions is limited
What do tell us such simulations ? • Theycorrelate pure possible processeswith patterns of evolution • Theycan not provethatsomeprocesscausedsomeevolutionaryresult, but theyprovidecandidate causal explanations: « if pattern X is met, then process x is likely to have produced it” • And other causal processes may have been at work but they were not so significant regarding such outcome (noise ???) • Even if we have no idea of the ecologicalcontext, hence of the actualselective pressures
Adami, Pennock, Ofria and Lenski (2003) show that evolution is likely to have favoured complexity : their point is that, if there is complexity increasing in their sense, then deleterious mutations might have been selected; then a decrease in fitness might have been involved in the stabilisation of more functionally complex genomes.
Chu and Adami (2000) investigation of the patterns of abundance of taxa : if the distribution of taxa resemble a certain power-law scheme X, it is likely that the parameter m (mean number of same order daughter families of a family) has been in nature close to the value of m involved in X (i.e. m=1).
The validation problem • Epstein (1999) : the case of Anasazi settlements
That does not prove that the rules ascribed to individuals are the accurate ones see also Reynold flocking boids : it excludes a centered-controlled social organisation (but we need other assumptions to make this plausible)
Even more : the case of Arakawa’s simulations in meteorology • Analysis by Kuippers & Lehnard 2001, Lehnard 2007 : drop « realism » in order to achieve efficiency
How are simulations to bevalidated in biology? • Mc Shea on complexity. Challenges Bonner (1988) explanation of the increase of complexitythroughselectedincerase of size in variouslineages Mc Shea (2005) suggeststhatcomplexityincreasecanbeproducedwith no naturalselection, only variation (complexitydefined by diversity); modelsalsoproduce patterns of complexityincreasewith patterns producedundervariousconstraints (driven trend vs passive trends, with no selection). The pattern found in the fossil records maybeproduced by suchprocess – but weneed to have an idea about the processeslikely to have actuallyoccurred
A minimal characterisation of computer simulations in evolutionarybiology : theyprovide candidate explanations (pure possible processes) and nullhypotheses for evolutionary patterns • For the samereason (theydon’taccept impure processeswhich are the onesreallyoccurring) theycan’tproveanything by themselves
An example worth to investigate : Hubbell’s ecological neutral theory (2001) • It skips the level of individual selection; generate the same outcome as what we see about succession, stability and persistance in communities
4.1. The longstanding problem with innovations Darwinism is gradualist (small mutations selected etc.) Cumulative selection accounts for adaptations
Novelties, • Innovations (qualitative, eg morphological, differences); • Key innovations : trigger adaptive radiation, and new phylogenetic patterns (avian wing, fish gills, language…) – id est, phylogenetic and ecological causal patterns
Pattern and processes : the role of « punctuated equilibria theory » (Eldredge and Gould 1976)
An issue with discontinuity • Problem : the fitness value of half a novelty ? (half a wing !) -> Solutions : • Find a benefit for each stage in variousspecies (Darwin on the eye) • Conceive of it as an exaptation (ex. feathers) (Gould and Vrba 1981) • Developmentalprocesses (Gould, 1977; Muller and Newman, 2005, etc.) : variation is not « minor », it’s a rearrangement of structures throughshuffling of developmental modules/time (as such the pucntuatedquilibria pattern don’trequire a specificprocess)
4.2. Exploringdiscontinuity: Compositional evolution (Watson 2005) : “evolutionary processes involving the combination of systems and susbsystems of semi-independently preadapted genetic material” (p.3). • consideration of building blocks obeying some new rules that are inspired by the biological phenomena of sex and of symbiosis proves that in those processes non gradual emergence of novelties is possible.
1. A system with weak interdepencies between parts can undergo linear evolution: increases in complexity are linear functions of the values of the variables describing the system. Algorithms looking for optimal solutions in this way are called “hill-climbers”; they are paradigmatically gradual. They easily evolve systems more complex in a quantitative way, but they can’t reach systems that would display innovations..
2. If you have arbitrary strong complexities between the parts, then evolving a new complex system will take exponential time (time increases as an exponential function of the number of variables). Here, the best algorithm to find optimal solutions is the random search. • 3. But if you have modular interdependencies (encapsulated parts, etc.) between parts, then evolving new complex systems is a polynomial function of the variables. (Watson 2005, 68-70)
Algorithms of the class “divide-and conquer” are dividing in subparts the optimisation issue, and divide in its turn each subpart in other subparts : the initial exponential complexity of the optimisation problem approached through random search is thereby divided each time that the general system is divided – so that in the end the problem has polynomial complexity. • Those algorithms illustrate how to evolve systems that are not gradual or linear improvements of extant systems; but as polynomial functions of the variables, they are feasible in finite time, unlike random search processes.
“Compositional evolution” concerns pure processes that embody those classes of algorithms with polynomial rates of complexification, and have genuine biological correspondents: sex; symbiosis. “mechanisms that encapsulate a group of simple entities into a complex entity” (Watson 2005, 3), and thus proceed exactly in the way algorithmically proper to polynomial-time complexity-increasing algorithms like “divide and conquer”.
Watson refined the usual crossover clause in GA, integrating various algorithmic devices (for ex. “messy GA”, according to Goldberg, Korb and Deb 1989) in order to account for selection on blocks that take into account correlation between distant blocks, hence creation of new blocks (Watson 2005, 77).
. This proves that processes formally structured like those encapsulated processes – such are symbiosis, endosymbiosis, may be lateral gene transfer – have been likely to provide evolvability towards the most complex innovations, the ones not reachable through gradual evolution
The bulk of the demonstration is the identity between algorithmic classes (hill-climbing, divide-and-conquer, random search) and evolutionary processes (gradual evolution, compositional evolution). • So the solution of the gradualism issue is neither a quest of non-darwinian explanation (“order for free”, etc.), nor a reassertion of the power of cumulative selection that needs to be more deeply investigated (Mayr *), but the formal designing of new modes of selective processes, of the reason of their differences, and of the differences between their evolutionary potentials. In this sense, discontinuities in evolution appear as the explananda of a variety of selective processes whose proper features and typical evolutionary patterns are demonstrated by computer sciences