Stellar Evolution theory is based upon
well understood physics and provides clear
predictions as to how a Colour-Magnitude Diagram
(CMD) will change due to effects of age and
metallicity. The theory has been tested by looking
at nearby coeval star clusters. The power of
applying CMD analysis to galaxies has been
demonstrated in studies of the Carina dSph
(Smecker-Hane et al. 1996). In Carina the
observation of separate, distinct Main Sequence
(MS) Turnoffs has forced us to believe that this
small, nearby companion of our Galaxy has had a
very complex star formation history. No similar
direct evidence for “bursting” behaviour on a
global scale has been seen indisputably in larger
systems. Partly this is due to the greater
distances of larger systems, but also to the
complications in distinguishing old star formation
events (> 1 Gyr old) in systems which are
currently forming stars. Where we lack MS turnoffs
we have to resort to statistical modeling of the
CMD. This has been applied using a a number of
different approaches, but all generally based on
Monte-Carlo techniques (e.g. Tosi et al. 1992;
Bertelli et al. 1992; Tolstoy &; Saha 1996).
However, even using these more sophisticated
analysis techniques, it is difficult to find
unique solutions. This is mostly due to the
age-metallicity degeneracy on the Red Giant Branch
(RGB). The RGB is usually the most populated,
easiest to observe phase of stellar evolution. The
Carina CMD reveals the dangers of blindly
interpreting the RGB, because from the RGB alone
it is impossible to extract the information
revealed by the MS Turnoffs.