DNA Mutation Rates
DNA Mutation Rates
DNA Mutation Rates
and
Evolution
Mitochondria are “organelles” within living cells that are responsible for making the
currency of energy called ATP (Adenosine Triphosphate), which all cells need to
function. Mitochondria carry their own separate DNA (mtDNA) that is independent of
the nuclear DNA of the same cell. Human mtDNA is composed of 37 genes totaling
about 16,000 base pairs. This mtDNA also mutates at a much faster rate than nuclear
DNA (nucDNA) does. Human mtDNA has been completely mapped and all the coding
regions are known (As well as the proteins or RNA for which they code). Some of the
mtDNA does not code for anything (thought to make these sections immune from
“natural selection pressure”), and are known as the "control regions". One particular
region appears to mutate faster than any other region (1.8 times faster), because the
variation among humans is greatest here.4 When the cell divides, each cell takes some
of the mitochondria with it. The mitochondria replicate themselves independently within
the cell. Beyond this, it has been generally assumed that mitochondria are always
passed on from the mother to the offspring without being involved with genetic shuffling
and recombining of mtDNA with the mtDNA of the father. Recently, however, this notion
has been challenged. As it turns out, many cases of paternally derived mtDNA have
been detected in modern families of humans as well as other species. Consider the
findings of an interesting study published by Schwartz and Vissing in the 2002 issue of
The underlying mechanism responsible for the elimination of sperm mtDNA in normal
embryos is not well understood. We speculate that the process in some cases may be
defective, allowing sperm mitochondria to survive and giving those with a selective
paternally inherited mtDNA have been detected by the polymerase chain reaction (PCR)
and of mouse oocytes microinjected with sperm support the hypothesis that sperm
mitochondria are targeted for destruction by nuclear-encoded proteins. We report the
case of a 28-year-old man with mitochondrial myopathy due to a novel 2-bp mtDNA
deletion in the ND2 gene (also known as MTND2), which encodes a subunit of the
mtDNA harboring the mutation was paternal in origin and accounted for 90 percent of
So, what does such a finding mean with regard to mtDNA mutation rates and
molecular clocks? Well, consider the following comments by Morris and Lightowlers
the mother…. Several recent papers, however, have suggested that elements of mtDNA
may sometimes be inherited from the father. This hypothesis is based on evidence that
mtDNA may undergo recombination. If this does occur, maternal mtDNA in the egg must
cross over with homologous sequences in a different DNA molecule; paternal mtDNA
seems the most likely candidate…. If mtDNA can recombine, irrespective of the
mechanism, there are important implications for mtDNA evolution and for phylogenetic
Before this evidence of paternal inheritance was discovered it was assumed that
mtDNA was strictly the result of maternal inheritance. Based on this assumption, it was
assumed that the mitochondrial offspring would get exact copies of the mitochondria
that the mother had except if there was a mutational error. This error rate in the non-
coding portion of mitochondrial DNA has long been thought to occur once every 300 to
The Berkeley biochemists who developed the theory, Allan Wilson, Rebecca Cann,
and Mark Stoneking, made several apparently reasonable assumptions. Since there
were no DNA changes due to genetic recombination events (ie: with paternal DNA - now
known to be a wrong assumption), they assumed that all changes in the mtDNA were
the result of mutations over time and that these mutations occurred at a constant rate.
On the basis of these assumptions, the researchers believed they had access to
something like a "molecular clock." Because mtDNA is thought to mutate faster than
nuclear DNA (nucDNA), it was thought that the faster mutation rate of mtDNA would
The original 1987 study involved mtDNA from 136 women from many parts of the
world having various racial backgrounds. The analysis seemed to support the idea of a
single ancestral mtDNA molecule from a woman living in sub-Saharan Africa about
200,000 years ago. Later, more detailed studies seemed to confirm this conclusion.
Unfortunately though, there was a undetected bias in the computer program as well as
with the researchers themselves. The researchers used a computer program designed
to reveal a "maximum parsimony" phylogeny or the family tree with the least number of
mutational changes. This was based on the assumption that evolution would have
taken the most direct and efficient path (which is not necessarily true, or even likely).
Also, the computer program was biased by the order of data entry to favor the
information entered first. This problem was recognized when the computer gave
different results depending on the order that the data was entered. Now, after thousands
of computer runs with the data entered randomly, it appears that the "African origin" for
modern humans does not hold a statistical significance over other possibilities.26
The problems with these studies were so bad that Henry Gee, a member of the
editorial staff for the journal, Nature, harshly described the studies as "garbage." After
calculated that the total number of potentially correct parsimonious trees is somewhere
that low-level mixing among early human populations may have scrambled the DNA
sequences sufficiently so that the question of the origin of modern humans and a date
for "Eve" can never be settled by mtDNA.22 In a letter to Science, Mark Stoneking (one
of the original researchers) acknowledged that the theory of an "African Eve" has been
invalidated.23
Another interesting aspect of the "molecular clock" theory is the way in which the
mutation rate itself was determined. Contrary to what many might think, the mutation
rate was not initially determined by any sort of direct analysis, but by supposed
phylogenic evolutionary relationships between humans and chimps. In other words, the
mutation rate was calculated based on the assumption that the theory in question was
already true. This is a rather circular assumption and as such all results that are based
on this assumption will be consistent with this assumption - like a self-fulfilling prophecy.
Since the rate was calculated based on previous assumptions of evolutionary time, then
the results will automatically "confirm" the previous assumptions. If one truly wishes
independent confirmation of a theory, then one cannot calibrate the confirmation test by
the theory, or any part of the theory, that is being tested. And yet, this is exactly what
was done by scientists such as Sarich, one of the pioneers of the molecular-clock idea.
Sarich began by calculating the mutation rates of various species "...whose divergence
could be reliably dated from fossils." He then applied that calibration to the chimpanzee-
human split, dating that split at from five to seven million years ago. Using Sarich's
mutation calibrations, Wilson and Cann applied them to their mtDNA studies, comparing
"...the ratio of mitochondrial DNA divergence among humans to that between humans
and chimpanzees."24 By this method, they calculated that the common ancestor of all
modern humans, the "African Eve", lived about 200,000 years ago.
Obviously then, these dates, calculated from the mtDNA analysis, must match the
presupposition. The circularity of this method is inconsistent with good scientific method
"mitochondrial clock" theory was and is basically a theory within a theory in that it has
that scientists did not catch this inherent flaw earlier. Interestingly enough though, this
flaw in reasoning was not detected for many years and perhaps would have remained
undetected for much longer if a more direct mutation-rate analysis had not been done.
Eventually, scientists, who study historical families and their genetic histories,
started questioning the mutation rates that were based on evolutionary phylogenetic
assumptions. These scientists were “stunned” to find that the mutation rate was in fact
much higher than previously thought. In fact it was about 20 times higher at around one
mutation every 25 to 40 generations (about 500 to 800 years for humans). It seems that
in this section of the control region, which has about 610 base pairs, humans typically
differ from one another by about 18 mutations. 3 By simple mathematics, it follows that
modern humans share a common ancestor some 300 generations back in time. If one
assumes a typical generation time of about 20 years, this places the date of the
common ancestor at around 6,000 years before present. But how could this be?!
Thomas Parsons seems just as mystified. Consider his following comments published
years, this corresponds to ~1/600 generations and an age for the mtDNA
molecular clock would result in an age of the mtDNA MRCA of only ~6,500
y.a., clearly incompatible with the known age of modern humans. Even
acknowledging that the MRCA of mtDNA may be younger than the MRCA
The calculation is done in the following way: Let us consider two randomly chosen
human beings. Assuming all human beings initially have identical mitochondrial DNA,
after 33 generations, two such random human families will probably differ by two
mutations, since there will be two separate lines of inheritance and probably one
mutation along each line. After 66 generations, two randomly chosen humans will differ
by about four mutations. After 100 generations, they will differ by about six mutations.
After 300 generations, they will differ by about 18 mutations, which is about the
observed value.
mitochondria to all living humans) lived about 100,000 to 200,000 years ago in Africa.1
The new calculations, based on the above experiments, would make her a relatively
young ~6,500 years old. Now, the previous notion that modern humans are up to
10,000 generations old has to be reevaluated or at least the mtDNA basis for that
More recent direct mtDNA mutation rate studies also seem to confirm the earlier
findings by Parsons and others. In an 2001 article published in the American Journal of
Human Genetics, Evelyne Heyer et. al., presented their findings of the mtDNA mutation
families than those based on phylogenetic comparisons. . . For the HVI sequences, we
obtained 220 generations or 6,600 years, and for the HVII sequences 275 generations
or 8,250 years. Although each of these values is associated with a large variance, they
both point to ~7,000-8,000 years and, therefore, to the early Neolithic as the time of
different, than the value of 6.3 reported in recent the recent pedigree study of
comparable size . . . In another study (Soodyall et al. 1997), no mutations were detected
transmissions by Howell et al. (1996), and nine substitutions were observed in 327
transmissions by Parsons et al. (1997). Combining all these data (1,729 transmissions)
results in the mutation rate of 15.5 (Cl 10.3-22.1). Taking into account only those from
2000; the present study) leads to the value of 7.9. The latter, by avoiding experimental
problems with heteroplasmy, may provide a more realistic approximation of the overall
44
mutation rate."
Also, consider an even more recent paper published in a 2003 issue of the Annals
of Human Genetics by B. Bonne-Tamir et al. where the authors presented their results
of a their study of "Maternal and Paternal Lineages" from a small isolated Samaritan
"Compared with the results obtained by others on mtDNA mutation rates, our
upper limit estimate of the mutation rate of 1/61 mutations per generation is in close
agreement with those previously published." 45 [compared with the rate determined by
by Denver et al. is also quite interesting. These scientists reported their work with the
mtDNA mutation rates of nematode worms and found that these worm's molecular
clocks actually run about "100 times faster than previously thought" [emphasis
added].46
"Extrapolating the results directly to humans is not possible, say the scientists. But
their results do support recent controversial studies suggesting that the human
molecular clock also runs 100 times faster than is usually thought. This may mean that
modern man, happened much more recently than currently believed, says the team.
'Our work appears to support human analyses, which have suggested a very high rate,'
says Kelley Thomas of the University of Missouri. 'This work is relevant to humans,'
says Doug Turnbill of the institute for Human Genetics and Newcastle University, UK. 'If
the human mutation rate is faster than thought, it would have a lot of impact in looking
sequences of DNA from people and other animals. 'This is kind of analysis that was
used to determine that the African origin of modern humans was about 200,000 years
ago,' says Thomas. 'The problem with this approach is that you are looking at both the
mutation rate and the effects of natural selection,' he says. The technique would also
miss multiple mutations in the same stretch of mtDNA, says Paul Sharp of the Institute
but share a female ancestor. This approach has revealed higher mtDNA mutation
rates. But the results have not been accepted by many scientists [emphasis
added].
Knowing the exact rate of mutation in humans is very important for forensic
science and studies of genetic disease, stresses Turnbill. Forensic identification often
rests on comparing samples of DNA with samples from suspected relatives. Faster
Obviously then, these rates, based on more direct observations, are nowhere near
things just a bit now doesn't it? Isn't it strange though that many scientists are still loath
to accept these results? The bias in favor of both evolution as well as millions of years
for assumed divergence times between creatures like apes and humans is so strong
that changing the minds of those who hold such positions may be pretty much
impossible.
There are many other potential problems for phylogenies that rely on mtDNA
sequence analysis and mutation rates. One problem is that mtDNA functions as a
single genetic locus, much like a single gene does in nucDNA. Studies that work off a
single genetic locus are more likely to be affected by random genetic changes than are
studies that include more than one locus (the more the better). Therefore, single locus
studies are less accurate in characterizing a population. Beyond this, the new evidence
may not be as valid as was previously hoped. Some nucleotide regions mutate slowly,
while others can mutate relatively rapidly.17 These mutational "hotspots" can mutate
fairly rapidly even within a single lifetime and are intuitively rather common in the
tissues and, unless they involve gametes, they are not passed on to the next
have tried to compensate for these problems, but the various methods have produced
with historical sequences often yield very difference results from those estimated by
indirect methods that are based on present day sequence differences. For another
historically sequenced penguins have shown that their mtDNA mutation rates are 2 to 7
times faster than had previously been assumed through indirect methods.20 Certain of
these problems have in fact led some scientists to stop using control-region sequences
Those scientist that continue to try and revise the molecular clock hypothesis have
tried to slow down the clock by showing that some mtDNA regions mutate much more
slowly than do other regions. The problem here is that such regions are obviously
affected by natural selection. In other words, they are not functionally neutral with
For example, real time experiments have shown that average mitochondrial
assumptions. And, this is the average rate of the entire mitochondrial genome of
16,000pb. One might reasonably think that all aspects of the hypervariable regions
(HVI & HVII) would have a higher than average rate of mutation if truly neutral with
regard to functional selection pressures. Given this, those "slowly mutating sites" with
rates as slow as 0.065 mut/site/myr (Heyer et al, 2001) would seem to be maintained in
Again, such non-neutral changes are not necessarily the reflection of elapsed time
since a common ancestor so much as they are the reflection of the different functional
As with mitochondrial DNA mutation rates, the mutation rates of nuclear DNA have
often been calculated based on evolutionary scenarios rather than on direct methods.
By such methods, the average mutation rate for eukaryotes in general is estimated to
be about 2.2 x 10-9 mutations per base pair per year.29 With a 20 year average
generation time for humans, this works out to be around 4.4 x 10-8 mutations per base
pair per generation. Since most estimates of the size of the diploid human genome run
around 6.3 billion base pairs, this mutation rate would give the average child around 277
mutational differences from his or her parents. This sounds like quite a high number
and it is in fact on the high end of the spectrum when compared to studies looking more
particular study by Nachman and Crowell estimated the average mutation rate
Using these sequence comparisons, "The average mutation rate was estimated to be
~2.5 x 10-8 mutations per nucleotide site or 175 mutations per diploid genome per
These estimates do actually seem reasonable since they seem to match the error
rates of DNA replication and repair that occur between the formation of a zygote in one
generation and the formation of a zygote in the next generation. In the illustration31
below, notice that from fertilization to the formation of a woman's first functional gamete,
it takes about 23 mitotic divisions. Men, on the other hand, contribute about twice as
many germ line mutations as women do.33 At least part of the reason is that their stem
cells keep dividing so that the older a man gets before having children more mitotic
divisions occur.
Now, consider that each diploid fertilized zygote contains around 6 billion base
pairs of DNA (~3 billion from each gamete/parent, using a conservative round
number).32 From cell division to cell division, the error rate for DNA polymerase
combined with other repair enzymes is about 1 mistake in 1 billion base pairs copied.42
At this rate, there are about 6 mistakes with each diploid cell replication event. With a
Of course, this is right in line with the mutation rates that are based on evolutionary
scenarios. However, some estimates place the overall mutation rate as low as 1
mistake in 10 billion base pairs copied.43 At this rate, one would expect around 0.6
mistakes with each replication event and only around 17 mutations per person per
generation. In any case, since the rate of 175 mutations per generation is more in line
with most of the current estimates for humans, this rate will be used as the basis for the
Such a high mutation rate might be a more of a problem than it is for humans if it
were not for the fact that much of the human genome does not seem to code for
anything and can therefore sustain mutations without significant detrimental effects on
the overall function of the organism. The amount of this non-coding DNA has been
estimated by calculating the coding portion of DNA and subtracting this from the total
genomic real estate. It seems as though the average coding portion of a human gene is
around 1,350 base pairs in size. Of course, this gene would code for a protein
averaging 450 amino acids.38 Now, multiplying this number by the total number of
genes should give a reasonable estimate of the coding genetic real estate.
However, there is some argument as to the total number of genes in the human
genome. For many years it was thought that humans had between 70,000 to 140,000
genes. However, scientists working on the human genome project made a surprising
discovery. When they finished the project in February of 2001, they estimated that the
actual gene count was somewhere between 30,000 to 40,000 genes.39 But a year later,
in February of 2002, at the annual meeting of the American Association for the
Velculescu, suggested that the real number of genes in the human genome may
actually be closer to 70,000 genes after all. He and his colleagues, at Johns Hopkins
University in Baltimore, Maryland, have gone back to the lab to look for genes that the
computer programs may have missed. Their technique, called serial analysis of gene
expression (SAGE), works by tracking RNA molecules back to their DNA sources. After
isolating RNA from various human tissues, the researchers copy it into DNA, from which
they cut out a kind of genetic bar code of 10 to 20 base pairs. Velculescu proposes that
the vast majority of these tags are unique to a single gene. If so, the tags can then be
compared to the human genome to find out if they match up with genes discovered by
the computer algorithms. Velculescu stated that only about half of the tags he used
match the genes identified earlier in the genome project. Therefore, he suggests that
The reason for the disparity may be that the standard computer programs were
largely developed for the genomes of simple (prokaryotic) organisms, not for the more
complex sequences found in the genomes of humans and other eukaryotes. "We're still
not very good at predicting genes in eukaryotes," said Claire Fraser of The Institute for
Genomic Research in Rockville, Maryland. “It is entirely possible that there could be
more than 32,000 genes, and SAGE is an important approach to finding them… You
absolutely have to go back into the lab and get away from the computer terminal." 40
So, it seems as though there is still some question as to exactly how many genes
the human genome contains. But, for the sake of argument, lets go with a lower
estimate of ~40,000 genes. With each gene averaging 1,350 base pairs in size, only
around 108 million base pairs out of 6 billion base pairs (diploid) would code for
anything. This is only around 1.8% of the total genome. Much of the rest of the human
genome (At least 50%) is composed of a large amount of “repetitive DNA” that is made
up of similar sequences occurring over and over.33,38 At least some of the other 48% of
the genome is thought to provide structural integrity as well as regulating the production
of the coding sequences of DNA as far as when, where, and how much protein to
not clearly understood. In any case, for the purposes of this discussion a rough figure
Since mutations are the only possible source of novel genomic function in the
evolution of living things, we should consider a few facts about these mutations.
Mutations are thought to be purely random events causes by errors of replication and
maintenance over time. They occur anywhere in the entire genome in a fairly random
fashion with each generation. Given this information, lets consider how these mutations
would build up and what effect, if any, they would have on a human lineage.
Newer research suggests a detrimental mutation rate (Ud) of 1 to 3 per person per
generation with at least some scientists (Nachmann and Crowell, 2000) favoring at least
1,000 to 1, it seems like the build up of detrimental mutations in a population might lead
Nachmann and Crowell detail the perplexing situation at hand in the following
probably higher.
The problem can be mitigated somewhat by soft
Nachmann and Crowell find the situation a very puzzling one. How does one get
rid of all the bad mutations faster than they are produced? Does their hypothesis of
“positive epistasis” adequately explain how detrimental mutations can be cleared faster
than they are added to a population? If the functional effects of mutations were
increased in a multiplicative instead of additive fashion, would fewer individuals die than
before? As noted above, even if every detrimental mutation caused the death of its
owner, the reproductive burden of the survivors would not diminish, but would remain
the same.
For example, lets say that all those with at least three detrimental mutations die
before reproducing. The population average would soon hover just above 3 deleterious
mutation rates. Over 95% of each subsequent generation would have 3 or more
deleterious mutations as compared with the original "neutral" population. The death
rate would increase dramatically. In order to keep up, the reproductive rates of those
surviving individuals would have to increase in proportion to the increased death rate.
The same thing would eventually happen if the death line were drawn at 100, 500,
1000, 10000 or more deleterious mutations. The only difference would be the length of
mutations in its gene pool beginning at a relatively "neutral" starting point. The
population might survive fairly well for many generations without having to resort to
huge increases in the reproduction rate. However, without getting rid of the
experiencing an exponential rise in its death rate as its average population crossed the
Since the theory of positive epistasis does not seem to help the situation much,
some other process must be found to explain how to preferentially get rid of detrimental
bunches. The new findings thus support the idea that sex
Well, the answer might be found in a combination of processes where both sexual
replication and natural selection play a role to keep a slowly reproducing population
from going extinct. For example consider the following chart showing how deleterious
while the numbers of those that have greater numbers of deleterious mutations build up
more and more. In this article Rice notes that in asexual populations the only way to
really overcome this buildup of detrimental mutations is to increase the reproductive rate
substantially. But, what about beneficial mutations? Rice comments, "Rare reverse and
against the flow of genetic polarization. But this is a minor influence, analogous to water
turbulence that occasionally transports a pebble a short distance upstream." 49 So, how
good and bad mutations. For example, lets say we have two individuals, each with 2
detrimental mutations. Given sexual recombination between these two individuals, there
is a decent chance that some of their offspring (1 chance in 32) will not have any
inherited detrimental mutations. But what happens when the rate of additional
To look into this just a bit more, consider another example of a steady state
population of 5,000 individuals each starting out with 7 detrimental mutations and an
reproductive rate of 4 offspring per each one of the 2,500 couples (10,000 offspring), in
one generation, how many offspring will have the same or fewer detrimental mutations
This Poisson approximation shows that out of 10,000 offspring, only 2,202 of them
would have the same or less than the original number of detrimental mutations of the
parent population. This leaves 7,798 with more detrimental mutations than the parent
selection must cull out 5,000 of these 10,000 offspring before they are able to
reproduce. Given a preference, those with more detrimental mutations will be less fit by
a certain degree and will be removed from the population before those that are more fit
(less detrimental mutations). Given strong selection pressure, the second generation
might be made up of ~2,200 more fit individuals and only ~2,800 less fit individuals with
the overall average showing a decline as compared with the original parent generation.
If selection pressure is strong, so that the majority of those with more than 7 detrimental
mutations are removed from the population, the next generation will only have about
reproductive rate of 4 per couple, only 4,400 offspring will be produced as compared to
10,000 originally. In order to keep up with this loss, the reproductive rate must be
increased or the population will head toward extinction. In fact, given a detrimental
this is about half that required for an asexual population (2eUd), it is still quite
significant.
In this light, consider that more recent estimates suggest that the deleterious
mutation rate is even higher. "extrapolations from studies of humans and Drosophila
(Mukai, 1979; Kondroshov, 1988; Crow, 1993) suggest that Ud > 5 is feasible." 49
However, the number of required offspring needed to compensate for such a high
detrimental mutation rate would soar to 148 per female per generation! The various
forms of positive epistasis (see illustration by Rice below) 49 do not solve this problem.
Also, what about the Y-chromosome in males? The Y-chromosome does not
undergo significant sexual recombination. Are the males of slowly reproducing species,
like humans, therefore headed for extinction at an even faster rate than females?
chromosome] can never be ‘repaired’ by recombination. This has led to suggestions that
the Y is destined for extinction – it will eventually dwindle to nothing. According to this
model, its role in sex determination will eventually be taken on by genes elsewhere in
50
the genome."
The author of the above quoted article goes onto point out that several species,
like the Armenian mole vole, are able to reproduce without the Y chromosome. While
this might explain where humans are headed, it doesn't seem quite clear as to just how
the Y-chromosome could have evolved over millions of years of time given its relative
. Harlen Bretz
Debates:
Ladder of Complexity
Evolving Bacteria
Irreducible Complexity
Crop Circles
Function Flexibility
Neandertal DNA
Human/Chimp phylogenies
Geology
Fish Fossils
Matters of Faith