Module 9.1.1: Large Scale Structure Observations and Redshift

Module 9.1.1: Large Scale Structure Observations and Redshift Surveys
So, we have seen how large-scale structure is supposed to form theoretically, but now, let us
take a look at what observations have to say.
[Slide 2] The basic paradigm, of course, is that density fluctuations from the early universe
collapse into structures that we see today and that formation continues through hierarchical
merging of smaller pieces into ever larger ones. So, clearly, we distinguish between galaxies
which are obviously seen as unit things in space and are dissipatively formed structures
achieving density contrast of the order of a million relative to the average density, which
clearly puts them apart from large-scale structure itself, consisting of assemblies of really
dark matter with galaxies used as tracers. So by large-scale structure, we talk about galaxy
groups, filaments, clusters, sheets, superclusters, and that is what we will be addressing now.
Note that the basic assumption here is that we do know that most of the mass is in form of
non-baryonic dark matter and that most of the baryonic matter we don't really see directly. So
we hope, and this turns out to be actually a pretty good assumption, that galaxies, visible
ones, trace the large-scale structure that's actually there in an invisible material.
The basic way of quantifying the large-scale structure is to first map it out, find out where all
the galaxies are and see what their density field is like. To do this, we obviously measure two
coordinates in the sky and we need a third, radial coordinate. This is obtained through
redshifts, assuming of course that Hubble expansion works. So, we need to do large redshift
surveys, measuring redshifts for as many galaxies as possible in order to trace the large-scale
structure.
It was recognized already in the early 20th century that galaxies are not randomly distributed
in space, but rather, they seem to congregate. However, it took until about 1970s to really
start quantifying this.
[Slide 3] So in 1930s, Shapley and Zwicky, who started the first systematic sky surveying,
noted that galaxies do seem to form large structures. Shown here on the left is one of the
contour maps of projected galaxy densities on the sky from one of Zwicky's atlases.
The quantitative analysis of galaxy clustering was started in 1950s by Shane and Wirtanen,
astronomers who counted galaxies per unit area on photographic sky plates taken at Lick
Observatory. Those so-called Lick galaxy counts turn out to be a basic material for a lot of
the early theoretical studies of large-scale structure. From 1950s to 1970s, more and more
attention was paid to it. Gerard de Vaucouleur in particular noticed what we now call the
Local Supercluster, and more and more galaxy redshift surveys were conducted. The redshift
surveys really came into prominence in the 1980s and beyond, and today we have redshifts
for maybe 2 million galaxies or so.
[Slide 4] Here is a simple projection on the sky of the 6,000 brightest galaxies. Two things
are immediately apparent. There is a belt where galaxies are absent and that is called the
zone of avoidance. We now understand that this is due to the interstellar dust absorption in
our own galaxy that we simply cannot see very much through the dusty interstellar medium
in the Milky Way's disk. Outside of that, you see that there is some structure.
Now, if you were to make a plot of the 6,000 brightest stars in the sky, it would be exactly the
opposite. They are all concentrated in what is the zone of avoidance of the galactic world, the
angular disk of the Milky Way, and otherwise, they'd be fairly randomly distributed, on the
rest of the sky.
[Slide 5] So here is the 3D picture of our immediate extragalactic neighborhood, the so-called
Local Group of galaxies. Milky Way and Andromeda are the two dominant galaxies in the
group and there is a large number of their dwarf galaxy satellites. Groups of galaxies are
indeed where most of the galaxies are. So they are the most common building block of larger
scale structures.
[Slide 6] We mentioned the Local Supercluster earlier. This was noted as an elongated
structure in the sky and that we are kind of in its plane. The center is on Virgo cluster, the
nearest cluster of galaxies to us and it's roughly 60 megaparsecs in scale. As superclusters go,
this is not a very large supercluster, but it is the one in which we belong.
In addition to the commonly used equatorial or ecliptic or galactic coordinates from
astronomy, we now define so-called supergalactic coordinates. The XY plane of the
supergalactic coordinates is the plane of the Local Supercluster, and we are at its origin, and
the vertical Z direction is orthogonal to it. Later on, we will see plots of some of the
distributions projected onto this XY plane of the supergalactic coordinates.
[Slide 7] So here is the rough 3D sketch of what the Local Supercluster might look like. This
plot is centered on the Local Group of galaxies, but really the center of the Local Supercluster
is at Virgo. The rest of the structures are various galaxy groups or individual galaxies.
Actually, there are very few isolated galaxies, most of them are in groups. It is not a
perfectly flat system like the disk of the Milky Way, but it is flat. Two other clusters that are
sort of peripheral to the Local Supercluster are Fornax and Eridanus. They are on sort of the
other side of the sky from Virgo.
[Slide 8] Now this becomes a little more apparent if we plot not 6,000 but 15,000 brightest
galaxies in the sky. This was typical of galaxy catalogs until say 1990s, where those are
galaxies selected by eye or on photographic sky survey plates. Here we can see that there is a
great circle shown here through these arches that does go through a lot of galaxies, and that is
indeed the plane of the Local Supercluster. To the upper left, sort of, you see a structure
called Centaurus or Hydra-Centaurus, this is another supercluster of galaxies, and in fact, it's
bigger than the Local Supercluster and we're falling into it. We'll talk more about that when
we talk about peculiar velocities. On the other side is Perseus-Pisces Supercluster, roughly
the same distance. So, superclusters themselves tend to form slightly more coherent
structures, and this is not surprising from everything that we know about large-scale
structure.
[Slide 9] So the mapping of large-scale structure begins with catalogs of galaxies on the sky
and those were counted by eye, well, originally by Messier and Herschel and so on. But,
really, the more modern ones begin with Shapley and Ames in 1930s and then the Lick
galaxy counts in 1950s, and there are famous catalogs like Uppsala General Catalog of
Galaxies and the Southern Extension. Still, the Shane-Wirtanen Lick counts were the largest
up until 1980s. They counted close to 1 million galaxies in the sky, which is why this was a
good statistical sample for the early studies. The next step came with the digitization of sky
survey plates, and that was done for both northern and southern skies, starting sometime in
the late 1980s. And APM is for Automated Plate Measuring machine in Cambridge, England
and people who did this work produced a catalog of roughly 2 million galaxies. The northern
version of it, from the Second Palomar Sky Survey, the so-called Digital Palomar
Observatory Sky Survey or DPOSS, counted about 50 million galaxies.
By the mid 2000s, Sloan Digital Sky Survey obtained much higher quality data for a couple
hundred million galaxies at first. Nowadays, they all have almost a billion, both stars and
galaxies. I n the future, we'll look towards even larger sky surveys and Large Synoptic
Survey Telescope or LSST, which certainly will not be completed in 2015, this is an old
slide, more like 2020, we'll probably map billions of galaxies.
But that's just pictures in the sky. In order to get the third dimension, redshifts are needed.
And the first modern redshift survey was probably the one done at the Harvard-Smithsonian
Center for Astrophysics in the mid 1980s. They obtained redshifts of the order of thousand
nearby galaxies and started to quantify large-scale structure near us. Along with its Southern
Extension, they covered in the order of 2,500 galaxies, this was already good enough to begin
first modern quantitative structures of galaxy clustering. Now you may recall the zone of
avoidance was placing this artificial belt of missing galaxies in the sky, simply due to the
extinction in our own Milk Way. The way to bypass this was to look at infrared, as infrared
light goes through the interstellar dust relatively unobscured.
By the mid 1990s there was a good catalog of infrared selected galaxies from the All Sky
Survey by IRAS satellite. That produced a fairly unbiased set of galaxies that enabled
astronomers to go deep into the zone of avoidance. The resulting redshift surveys now went
up to 9,000 galaxies. Encouraged by their initial success, the Center for Astrophysics
conducted a second generation redshift survey which compiled the redshifts of about 25,000
galaxies. John Huchra was the leading astronomer in that enterprise, and also, his
collaborators including Margaret Geller and others. Roughly at the same time, Carnegie
Observatories did Las Campanas Redshift Survey in the southern hemisphere using one of
the first modern multi-object spectrographs, and they obtained redshifts for a comparable
number, about 23,000 galaxies. These surveys of 1990s started revealing more interesting
features in the galaxy distribution, but really, the state of the art was achieved by two large
surveys, one in Australia, the two-degree field (2dF) redshift survey, and the Sloan Digital
Sky Survey, we'll talk a little more about them later, which together, now mapped well over a
million galaxies.
[Slide 10] Let's now look at the sky in the successive redshift shells as we step out. The
closest one would be shown here. These are galaxies with the recession velocities due to the
Hubble Expansion up to 3,000 kilometers per second, and the biggest feature you see is the
Local Supercluster. Most of the galaxies are being centered on Virgo, but with some pieces
in Fornax. Stepping further out between 3,000 and 6,000 kilometers per second, new
structures come in, the most notable one is the Perseus-Pisces supercluster opposite from
Virgo. This was largely done by Giovanelli and Haynes at Cornell and their collaborators,
who conducted their work in parallel with the CfA group. And again, sort of opposite to
Perseus-Pisces and almost hidden by the Galactic plane, is the Hydra-Centaurus Supercluster,
which later made an appearance under the name of Great Attractor, but we'll talk about that
later. And now, even further out between 6,000 and 9,000 kilometers per second, the biggest
feature is the Coma - Abell 1367 supercluster, composed largely of these two rich clusters.
But there are also some vestiges of others, especially the Perseus-Pisces.
[Slide 11] And so here is a 3D sketch of what our local supergalactic neighbourhood might
look like, with the Local Supercluster and nearby superclusters like Perseus-Pisces and
Hydra-Centaurus. You have probably noticed that superclusters, at least those near us, are
given names by the constellations in which most of the galaxies are placed. Constellations of
course have no physical meaning, but they're still used as a traditional convenience to
designate where in the sky some things might be.
Module 9.1.2: Large Scale Structure Observations and Redshift Surveys, cont.
[Slides 1-2] So, once you get a 3-D distribution, you can project it onto the supergalactic
plane. And here is a density landscape, high peaks corresponding to high projected density
from one of the IRAS-based redshift surveys. So the big set of mountains off from the center
is the Local Supercluster. We are always in the middle, so we're sort of on the slopes of it.
[Slide 3] But why project on the supergalactic plane if you have 3-D picture? And so, here is
a highly smoothed density distribution of galaxies in 3-D from the IRAS Redshift Survey.
Now you see that there are really all these blobs that really are in complex relation to each
other.
[Slide 4] I mentioned a second Center for Astrophysics survey by Huchra and collaborators,
and what they did is they chose a slice in the sky, very thin in declination, long in right
ascension, fan-shaped like that. They did this simply in order to have manageable numbers
of galaxies to observe, they went deeper than the previous survey, and went past the Coma
Cluster of galaxies, and obtained this famous Stickman Diagram. The structure in the middle
is the Coma cluster, and it actually doesn't look like that, it is elongated in radial direction by
the so-called finger of God effect. Namely, remember, that measured velocities of galaxies
are the vector sum of the Hubble expansion velocity and a peculiar velocity the galaxy may
have on its own. Now, in rich clusters of galaxies, galaxies have high velocity dispersions.
They have to move fast in order to balance the gravitational potential of the dark matter. So,
if you look through a cluster of galaxies, suddenly some galaxies will appear to be closer to
us because they're moving towards us relative to Hubble expansion. Some are moving away
from us, and there will be an elongation in what we call the redshift space, as opposed to
actual 3-D space.
So, this is a two-dimensional slice through a three-dimensional structure. And you can see
that galaxies actually form a kind of mesh of filaments, and there are big voids that had been
noticed previous to this. Because this is a two-dimensional cut, then it's reasonable to assume
that, in fact, voids must be really more of a spherical nature, or at any rate, not flat, and that
sheets intersecting through this plane may actually look like filaments. In reality, there are
both sheets and filaments out in the large-scale structure.
[Slide 5] So, there are actually two important effects about the redshift space. And remember,
redshift space has the radial coordinate that is associated with total measure velocity of the
galaxy. And a priori, we do not know which part is due to Hubble velocity, which would be
the real radial coordinate, and which part is due to peculiar motion. Statistically, we can
decompose that later. Then what happens, in rich clusters of galaxies, as I already
mentioned, there will be an elongation towards the observer, due to the radial velocity
dispersion of galaxies moving fast inside a cluster. On the other hand, if we're looking not at
a dense structure like cluster but, say, an intersection with one of the sheets or walls or
filaments, galaxies will be falling towards it. It's not yet a virialized structure, but it's a
density excess, and so galaxies fall to it. So, those which are on our side are falling towards
it, acquiring velocity that makes them look like they're a little further out. Those on the other
side are falling back towards that filament and that will make them look like they're a little
closer. So then, this will squish the structure along the line of sight, the opposite of the finger
of God effect. It all depends on how dense or how virialized is the density that we see. In
cases of higher densities like clusters of galaxies, we see the elongation, the finger of God
effect. In cases of low density enhancements, like filaments, we see the opposite. So, these
effects are well-understood and they have to be deconvolved out of the observations when
performing the full analysis.
[Slide 6] So, the full compilation of CfA2 redshift survey by Huchra et al. went well past
10,000 kilometers per second and they noticed that there is an even larger structure they
called the Great Wall, which contains, in part, the Coma - Abell 1367 cluster, or maybe
supercluster. And it was basically filling up the space available to their survey. Up until
then, every time a redshift survey was done, a structure was seen that was as large as it can be
fit in the survey volume. Which, on face value of it, doesn't bode well for the assumptions of
homogeneity. The resolution of this came with Las Campanas redshift survey, but before I
tell you about that, let me just give you some idea of how these measurements are done.
[Slide 7] In the early days, spectrographs were doing one galaxy at a time. This was a tedious
work and that's what CfA survey was. But then, astronomers designed multi-object
spectrographs, and they come in two varieties. First, we have to know, of course, where the
galaxies are in the sky. In one type, you have a metal plate that's positioned in the focal plane
of a telescope, and a little opening or slit was made where each galaxy should be. Then,
instead of having one spectrum with lot of blank space on each side, you have a whole lot of
little short spectra with object in the middle. The other way is to use optical fibers. Again,
the fiber head is positioned to where the galaxy should be, using some kind of device, usually
a robotic arm. And then, these fibers are grouped together to enter the slit of the spectograph,
and then the spectra are taken together. Both approaches worked fairly well. And this is what
really moved the redshift surveys to industrial strength of hundreds of thousands and now
millions.
[Slide 8] Recall that the second CfA survey was a slice in the sky. A fan-shaped slice, thin
declination, long in right ascension, and they saw all these voids and filaments and stuff. Las
Campanas Redshift Survey did the same thing but they did three slices, and they also went
further out. So, in their survey, they could see structures of same size up to a hundred
megaparsec or so, that were seen by the CfA Redshift Survey as well. But they also went
further, and no larger structures were seen. In other words, homogeneity does seem to apply
on scales larger than hundreds of megaparsecs. And for cosmological purposes, that's
perfectly good enough.
[Slide 9] Two huge redshift surveys really transformed this field. The first one was done by
Anglo-Australian observatory, an Australian-UK Consortium, who used one of those robotic
fiber spectrographs on the 4 meter Anglo-Australian telescope in Siding Spring in Australia.
They collected about a quarter million redshifts and they did the job first. They also observed
a lot of quasars, and so on. Their analysis really started revealing interesting new features of
large-scale structure. At the same time and continuing beyond, was the Sloan Digital Sky
Survey. They used a 2.5 meter telescope in New Mexico, at Apache Point Observatory. And
they did both imaging survey and spectroscopic survey. They did imaging of large areas on
the sky first, selected galaxies, then they used multifiber spectrographs to obtain redshifts of
galaxies, as well as quasar,s and so on. This was spectacularly successful, an earlier release
of the Sloan Digital Sky Survey had something like 800,000 galaxy redshifts, but they did
continue since then. And now, they have of the order of 1 and a half million galaxy redshifts,
and a comparable number of velocity measurements for stars, as well as a couple hundred
thousand quasars. This is truly a fundamental data set on which many of the modern studies
of the large-scale structure are based.
[Slide 10] So, here is the redshift slice projection of the 2dF survey. And it looks sort of like
Las Campanas Redshift Survey, only with much better resolution, because they had an order
of magnitude more galaxies. You can see the mesh of filaments and voids and so on, but you
can also see that on scales larger than about 100 megaparsecs, we do not see super voids or
super filaments or anything like that.
[Slide 11] Here is a picture showing sky coverage from the latest incarnation of Sloan Digital
Sky Survey. They have these strange belts in the sky because they operate in drift scanning
mode. Telescope moves in large circles in the sky as they collect the data. And then they
obtain redshifts along the same areas. This may look a little strange at first. But if you know
exactly where you are pointing, you can certainly take this into account in your analysis.
[Slide 12] And here is the projection of the Sloan Sky Survey. Now again, this squishes 3D
information into a plane that cuts across the sky. And we see the same qualitative features as
before: a frothy kind of large-scale structure, composed of filaments or walls or sheets
intersecting and more or less quasi-spherical voids inside of them, sort of like sponge-like
topology.
[Slide 13] There is always a tradeoff, of covering a large area on the sky, and being complete,
versus going deep. If you have a finite amount of observing time, there are only so many
galaxies you can observe. And so, you can trade one for the other. These large scale surveys
that we described covered large areas, but, they did not go very deep on cosmological scales.
Complementary to them were surveys done over small areas in the sky, but going very deep,
so-called pencil beam surveys, because of the narrow cone. The first of those were done in
1990s, and they saw presence of large-scale structure out of high redshifts, spikes in the
redshift distribution. At first, it seemed like those are periodic but that turn out not to be the
case and now it's believed that the characteristic scale that they saw there was really due to
the baryonic acoustic oscillations at scales of 100 megaparsecs or a little more.
[Slide 14] Here is the little redshift slice distribution from one of these modern survey. This
one is called DEEP and it was done at Keck telescope reaching out to redshift 1.5, or several
times deeper than the large area surveys from before. Now, if you make a histogram of
redshifts along the line of sight, you see these spikes. They correspond to intersections of
clusters and filaments and walls along your line of sight. And the interesting thing is that
structure keeps persisting out to the largest distances we can measure. This leads to an
interesting new phenomenon called biasing, which we'll discuss in a little more detail.
[Slide 15] So, here is a comparison of some of the redshift surveys. There are many ways in
which you can compare redshift surveys. In this particular case, it is the number of objects
measured versus the volume of space that's been covered. Obviously, the deeper you go, the
more volume you get at the expense of taking more spectra, having more objects. Now, as the
SDSS-Main is the actual Sloan Redshift Survey, they had few others. For one of them, they
didn't really actually use spectroscopic redshifts. They used multicolor photometry to derive
statistical estimates of redshifts, the so-called photo-z survey. This turns out to work
reasonably well if you have good photometry and at least several filters. And also they can
look at spectra of quasars, very far away, and look at absorption line clouds along the line of
sight. Those are very numerous and they provide means of probing large-scale structure out
to the highest redshifts.
[Slide 16] Another way to compare them is to look at area coverage versus number density of
objects that are being covered. And so again, there is a trade-off. Now, in this plane, diagonal
lines show the total numbers of objects, as you can see progressing first from thousands to
now millions.
[Slide 17] Next time, we will talk about how we actually quantify galaxy clustering using the
two-point correlation functions.
Module 9.2: Galaxy Clustering: The Two-Point Correlation Function
We have seen how redshift surveys can be used to describe large-scale structure, at least in a
cosmographic sense. But how do we quantify the distributional galaxies such that we can
compare with theoretical models? The first way in which people have done this is to use the
so called two-point correlation function. Nowadays, power spectrum is more often used and
the two are actually related in a very simple fashion. We will go about that next time. So,
let's talk about the galaxy two-point correlation function.
[slide2] What it means is that, if galaxies are clustered together, they're correlated. Each
galaxy is somehow more likely to be found next to another galaxy. And one way to quantify
this is to ask this question:
Assume that galaxies are actually uniformly, randomly distributed in space. Then, at a
distance from any given galaxy, there will be a certain probability of finding another galaxy.
If you actually measure this, you'll find out, there is an excess: there are more galaxies near
other galaxies than you'd expect from purely [uniform] random distribution. And that excess
above the random is what correlation function is.
So, one simply does the counting of galaxy pairs for each galaxy and then normalizes by
what the random distribution with the same number of data points would be. As it turns out,
the two-point correlation function is well represented by power-law, and it's usually written
in this form. That, radius, divided by some scaling radius, to some power, gamma, which is
close to -1.8.
And typically, for normal galaxies in this neck of the universe, the scaling length is about 5
megaparsec. This, however, is not universal. Different kinds of galaxies have different
clustering properties, as you'll see shortly.
[slide 3] So, here is an example of a modern, well-measured, two-point correlation function
from the galaxies from 2dF redshift survey. It does look pretty close to the power-law.
Although, if you subtract the best fit power-law, you'll see that there are some significant
deviations from it.
[slide 4] Now, if you don't have redshifts, you can measure the angular correlation function
just projected in the sky. And that this two-dimensional projected correlation function,
usually denoted with a little w, not to be confused with the equation of state parameter, is
related to the three-dimensional correlation function, xi of r, in a fairly simple fashion. The
power-law exponents differ exactly by one because we reduced the dimensionality of
problem from 3 to 2.
Another important point is that if galaxies are more likely to be found near other galaxies,
there is that excess probability. Then, in order to keep the average constant, it has to turn
negative at some point, and it does, at the scales that are roughly corresponding to those of
voids seen in galaxy distribution. If there is a void in distribution, in some sense, that's anticorrelation. It's less likely to find galaxy there, than if it would be, if the space was uniformly
populated with galaxies.
[slide 5] So, how do we do this in practice? Suppose we have a catalog of galaxies from some
survey. Then, you can create a random catalog of galaxies, with the same number of
galaxies, but randomly distributed in a Poissonian fashion. Then, you can do the simple
counts. Count galaxy pairs, one against the other, and count the fake galaxies pairs, from the
random catalog, divide the two and subtract one because it is the excess probability. For
large extensive catalogs of galaxies with uniform borders, this will work fairly well. But in
reality, catalogs do have some incompleteness or uneven borders, and so on. So, there is a
better estimator, which is called the Landy-Szalay estimator, and its formula is given here.
We make a correction by counting galaxy versus random catalog pairs. This takes care of the
boundary conditions.
[slide 6] Now, if you're not counting the individual galaxies, but have, essentially, a galaxy
density field where you can divide galaxies in boxes or pixels, then numerical galaxy density
can be used in the same fashion. Just subtract the expected average from the actual count n,
and divide by the average. So, if you do this, for any kinds of pairs or density pairs, then xi
of r is really the expectation value of this probabilistic distribution.
Now, strictly speaking, because we are correlating galaxies with themselves, this should be
called the autocorrelation function. But common usage is just to call it correlation function.
Note, also, that you can correlate sample of one kind of objects versus the others. So, for
example, you could ask, are galaxies clustered around quasars? And we can evaluate the
cross-correlation function between galaxies and quasars, and so on.
By analogy, we can expand this to higher-order terms. The three-point correlation function,
now asks, given a galaxy, given a probability of finding another galaxy at certain distance,
what is now probability of finding a third galaxy at some other distance? This obviously gets
a lot more complicated and numerically tedious very fast. However, there is some useful
information in these high-order correlation functions, and sometimes they're evaluated.
[slide 7] I mentioned that different kinds of galaxies can cluster differently, and here is a
simple thing you can do. You can ask, how are galaxies clustered, say, bright ones versus
faint ones? And it turns out that bright ones are clustered more strongly. The amplitude is
higher and the slope is steeper. The steepness of the slope obviously correlates with how
strongly they are clustered. If the correlation function was perfectly flat, there would be no
correlation. Understanding effects like this contains some useful information about galaxy
formation mechanisms. And we'll talk more about those later in the class.
[slide 8] Or, you can divide galaxies by morphological type, say, ellipticals versus spirals. It
turns out that elliptical galaxies or redder galaxies are clustered more strongly than the blue
ones, the disk galaxies. That, too, has some interesting clues about galaxy formation and
evolution.
[slide 9] So, it's a power-law. Does that mean it's a fractal? Remember, for any distribution
of points, the probability of finding another one from the same set increases the sum
dimensionality of the space. In normal 3-D space, this would be like cube of radius. If that
number is not an integer, the set is called fractal. So, we can write this formula, which is
resembling correlation function. And if, indeed, it was pure power law, then you could say,
universe was fractal, if it was a pure power-law extending to infinitely large distances. But in
reality, that's not the case. There are significant deviations from a power-law, it is slightly
bent. And, therefore, universe is not fractal or large-scale structure is not fractal, although it
is pretty close to it.
[slide 10] Next time, we'll talk about power spectrum of galaxy clustering, which can be
directly related to theoretical predictions.
Module 9.3: Large Scale Structure: Power Spectrum
[slide 1] Finally let us address the subject of the power spectrum of large-scale structure. You
may recall that we introduced this concept when we talked about primordial density field.
And now, just like then, some familiarity with Fourier analysis is absolutely essential. So if
you are familiar with Fourier analysis, the main subject of today's lecture will be almost
trivial. And if you are not, it will make no sense whatsoever. So this would be like really
good time to refresh your knowledge of Fourier analysis.
Done? Okay, well that was fast. So let's get going.
[slide 2] So we can characterize the density field of large-scale structure in the same ways we
did before. So let's take this Fourier transform. Here is the formula. It's essentially almost
the definition of Fourier transforms. And, the inverse transform is given by this. So in this
way spatial scales and wave number scale are connected, and the power spectrum of density
fluctuations is obviously a complex product of spectrum by itself.
Now recall what was the value of 2-point correlation function expressed from density field of
the expectation value. In other words, this is exactly what it is. So, correlation function is a
transform of the power spectrum. Power spectrum and correlation function are a Fourier pair.
They are completely equivalent to each other mathematically.
[slide 3] And here is an example for what they look like from Las Campanas redshift survey,
that you may recall. What is shown here is both three-dimensional 2-point correlations
function and power spectrum corresponding to the same large scale structure. You may
recall that 2-point correlation function was fairly easy to evaluate in any one of several ways.
And, but, power spectrum is really what theory produces. So, one can be translated into the
other.
[slide 4] Now, we usually express the power spectrum on a log-log axis, and theoretical
models, fitted against observations, give us the shape of it. However, the amplitude has not
been defined. There are two ways in which that can be done. One is to measure fluctuations
so that the density field is at a particular scale. And typically, sphere of radius of 8 h-1
megaparsecs is used because that gives the RMS of fluctuations close to unity. Turns out it's
a little less than that for modern measurements, but traditionally the scale of 8 megaparsecs is
used. So, whatever the value of that is, gives us unit normalization of the power spectrum at
that spatial scale.
Essentially what's been done here is you're convolving what's technically called a spherical
top-hat filter, but it's really nothing but a sphere, with a power spectrum itself. An alternative
way is to simply use observations of cosmic microwave background, which give us power
spectrum on very large scales. So if you have a model like cold dark matter model, you can
fit it on those scales, evolve for the appropriate linear growth of fluctuations, and that works
just as well.
[slide 5] So, here is one of the modern renderings of the power spectrum of large-scale
structures. Spanning the range all the way from super horizon sizes, down to the scales of
galaxies. The plot assembles several different kinds of measurements. They are all brought
to the same redshift by scaling in the linear regime. At large scales, cosmic microwave
background is what's being used. That overlaps with measurements of clustering from large
scale structure from redshift surveys and such. And then going to smaller scales, we used
things like abundance of clusters of galaxies, that's essentially statistics of high density peaks,
and clustering of Lyman-alpha clouds, which are sub-galactic fragments. Weak gravitational
lensing reflects the density distribution of the dark matter, and so that too can be used to
probe the distribution of CDM on intermediate scales.
And the line going through at points is the standard cold dark matter model. As you can see,
it's a remarkably good fit. And also an excellent agreement between these completely
different ways of observing it in the regions where they overlap. So this gives us a real
confidence that we actually know at some level of certainty what's going on, and one of the
major reasons why people believe CDM model is correct one.
[slide 6] Just as an aside, you may recall the baryonic acoustic oscillations. And, this is
actually how they were detected. By looking at the 2-point correlation function or power
spectrum at scales in the vicinity of hundred megaparsecs. So this is an absolutely essential
use of this concept as a cosmological test that can be used to improve our knowledge of
cosmological parameters.
[slide 7] So far we talked about power spectrum, that is the distribution of amplitudes of
density fluctuations. But in that process, as in computing power spectrum, we completely
neglect phase information.
Now here is a dramatic illustration of why phase information is important, produced by Alex
Szalay. And it's a toy model, the density model that is really Voronoi foam, kind of
mimicking the filamentary structure that we see in the large-scale structure. What was done
here, is that its Fourier transform was taken, and then the phases were randomly scrambled,
leaving the power spectrum identical, then transformed back, with the scrambled phases, and
you see the plot on the bottom. Looks like a complete noise. The coherence is completely
gone. So, whereas the distribution of fluctuations is the same, their arrangement is
completely different. The phase coherence has been lost in this computation, and obviously
we have to somehow learn how to use statistics of the phases of the density field power
spectrum or rather the Fourier spectrum in order to characterize it fully. Because when you
look at large scale structure, you don't see just lumpy density field. You do see a mesh of
filaments and voids, and what not. Amazingly enough, so far nobody has come up with a
really simple, elegant, complete way of describing this, so maybe you will.
[slide 8] So we discussed clustering of galaxies at some detail, but what about clustering of
clusters of galaxies? Or in fact you can discuss clustering of any elements of large scale
structure galaxies, groups, clusters, super clusters, and so on.
As it turns out, clusters also cluster and their clustering is stronger than that of galaxies.
Actually, there is an excellent correlation between the strength of clustering, say measured
through clustering length, and the richness of the system at which you're looking. The richer
clusters are clustered more strongly. This is actually a very interesting result, and at first it
was puzzling why should that be so, until Nick Keiser came up with the explanation that this
is due to phenomenon now called bias.
[slide 9] You can consider density field, and so there are some high peaks and low peaks.
Remember it can be also be composed in waves of large wavelength and small wavelength
fluctuations riding on top of large waves. So, if you impose a threshold for a formation of
certain kind of object, say, 5-sigma cut, then the highest peaks will be strongly clustered,
because the smaller fluctuations are carried together by the large waves, and those that
happen to be a crest of a big wave will be all clumped together.
This is in same sense as if you were to ask what are the highest points on the surface of planet
Earth. And if you put a cut at few kilometers elevation above the sea level, you'll see they're
very strongly clustered where the high mountains are, in this case, say, Himalayas and
Cordilleras and so on. Whereas if you lower down the threshold to the ground level, then
everything will be more or less uniformly distributed.
So this is a simple explanation. The highest density peaks always cluster more strongly than
the lower density peaks. And this applies to galaxies as well. We find out the denser galaxies,
elliptical galaxies, say, cluster more strongly than spiral galaxies. And I've shown you plot of
that we've seen earlier. This is the explanation why. It also has some implications. The
highest density peaks will be the first ones to form at small scales. This, we'll come back to
this phenomenon when we talk about galaxy formation and evolution.
[slide 10] So let us recap everything we learned about large scale structure. So we see a range
of scales from galaxies, which are scales of kiloparsecs, through groups and clusters, on
scales of megaparsecs, to superclusters on scales of maybe hundred megaparsecs. We
measure it, largely, by using redshift surveys of galaxies, and today we have redshifts for
couple of million galaxies, which gives us really excellent insights into what large scale
structure looks like as a function of redshift.
But clustering itself is not enough. There is also lot of interesting topology going on.
Coherent structures like filaments and voids and sheets that separate them, and so on. We
understand why this happens. You may recall that an aspherical fluctuation will first turn
around and collapse along one axis while expanding in the other two, making things like
sheets or walls, then will turn around on intermediate axis, collapse into a filament-like
structure, and eventually all drain into, again, to quasi-spherical one. o this is the origin of
the large scale topology, although we still don't have good ways of characterizing it.
Clustering itself is customarily characterized through 2-point correlation function or
equivalently the power spectrum. The 2-point correlation function has a fairly robust
functional form. It is well represented, not perfectly, but well, over large range of scales as a
power law with a slope of -1.8 in three-dimensional space. For galaxies, the normalization
given through clustering length is of the order of five megaparsec or so. For clusters, it's
larger, responding to larger amplitude of clustering. And whereas we can easily measure a 2point correlation function for any kind of objects, to compare them with theory, we actually
have to turn 2-point correlation function into a power spectrum. And fortunately the two are
equivalent, and therefore a pair, so we know how to compute that.
Overall, the cold dark matter model fits the data remarkably well over a full range of scales
that we can measure through a variety of different methods, which gives us confidence that
cold dark matter model is, in fact, correct. And I should note at this point, that would be very
difficult to achieve this if there were no dark matter at all. Generally speaking, objects of
different kinds have different clustering strengths that applies to different kinds of galaxies:
luminous versus less luminous, early versus late types, and so on, carrying over to the clusters
of different regions, and so on.
[slide 11] Next, we will talk about large-scale peculiar velocity field, which is a direct
consequence of the existence of large-scale density field.