Measurement of BAO correlations at with SDSS DR12 LyForests
Key Words.:
cosmology, Ly forest, large scale structure, dark energyWe have used fluxtransmission correlations in Ly forests to measure the imprint of baryon acoustic oscillations (BAO). The study uses spectra of 157,783 quasars in the redshift range from the Sloan Digital Sky Survey (SDSS) Data Release 12 (DR12). Besides the statistical improvements on our previous studies using SDSS DR9 and DR11, we have implemented numerous improvements in the analysis procedure, allowing us to construct a physical model of the correlation function and to investigate potential systematic errors in the determination of the BAO peak position. The Hubble distance, , relative to the sound horizon is . The bestdetermined combination of comoving angulardiameter distance, , and the Hubble distance is found to be . This value is times the prediction of the flatCDM model consistent with the cosmic microwave background (CMB) anisotropy spectrum. The errors include marginalization over the effects of unidentified highdensity absorption systems and fluctuations in ultraviolet ionizing radiation. Independently of the CMB measurements, the combination of our results and other BAO observations determine the openCDM density parameters to be , and .
1 Introduction
The sound waves that propagated in the prerecombination universe produced a pronounced peak in the twopoint correlation function of the cosmologicaldensity field (Peebles & Yu, 1970; Sunyaev & Zeldovich, 1970). This “baryonacoustic oscillation” (BAO) peak, first observed by Eisenstein et al. (2005) and Cole et al. (2005), is centered on a comoving distance equal to the sound horizon at the drag epoch, . Observed at a redshift , the BAO peak position in the transverse (angular) direction determines the ratio , where is the “comoving” angulardiameter distance ( is the “traditional” angulardiameter distance). In the radial (redshift) direction, the peak position determines , where is the Hubble distance. Since , and all have simple dependencies on the cosmological parameters, observation of the BAO feature constrains those parameters, especially when combined with cosmic microwave background (CMB) data (Planck Collaboration et al., 2014, 2016; Aubourg et al., 2015). In particular, one can constrain models beyond the flatCDM model that describes CMB data, deriving constraints on cosmological curvature and the darkenergy equation of state. Furthermore, the shape of the spectrum of CMB anisotropies can be used to calculate the value of to percentlevel precision, Mpc and the use of this value allows one to derive and from BAO measurements. These absolute distances can be combined with relative distances determined with type Ia supernovae (Betoule et al., 2014), to extrapolate to , yielding a “topdown” measurement of (Aubourg et al., 2015).
Most studies of the BAO peak have used galaxies at redshifts as tracers of the density. The first observations used data from the Sloan Digital Sky Survey (SDSS) (Eisenstein et al., 2005) and data from the TwoDegree Field Redshift Survey (Cole et al., 2005), and the combination of these sets (Percival et al., 2007, 2010). Since then, results of increasingly higher precision have been obtained, most significantly in the redshift range from the Baryon Oscillation Spectroscopy Survey (BOSS) of SDSSIII (Anderson et al., 2012, 2014b, 2014a) with the results of the complete survey being summarized in Alam et al. (2016). Results at other redshifts have been obtained by 6dFGRS at (Beutler et al., 2011), WiggleZ at (Blake et al., 2011), SDSSI at (Padmanabhan et al., 2012; Mehta et al., 2012; Chuang & Wang, 2012; Xu et al., 2013) and SDSSI at (Ross et al., 2015). There is an impressive agreement of the results of these studies with the expectations of flatCDM models based on CMB data, as emphasized by Planck Collaboration et al. (2016).
At higher redshifts, BAO correlations can be seen using quasars and their Lyman (Ly) forests as mass tracers (McDonald & Eisenstein, 2007). The correlations in the Lyforest fluxtransmission field of BOSS quasars were first studied in Slosar et al. (2011) and the BAO peak was seen in SDSS data release DR9 (Busca et al., 2013; Slosar et al., 2013; Kirkby et al., 2013) and DR11 (Delubac et al., 2015). Crosscorrelations of the Ly absorption with the distribution of quasars were detected in DR9 (FontRibera et al., 2013), and the first BAO detection was presented in FontRibera et al. (2014).
In this paper, we study the autocorrelation function, , of the Ly fluxtransmission field using the SDSS data release DR12 and we update the cosmological constraints reported in Delubac et al. (2015). Our study of the quasarforest crosscorrelation function will be presented in a future publication (du Mas des Bourboux et al., in preparation). In addition to using the increase of survey area of DR12 over DR11, the following improvements over the analysis of Delubac et al. (2015) have been implemented:

Spectroscopic pipeline improvements (Sect. 2) that includes a new algorithm for the extraction of spectra from the CCD images that results in a more linear flux response. This modification allows us to correct for the mean distortion of the fluxtransmission field due to imperfect spectral modeling of standard stars. We also correct for the differential positioning of quasar and stellar fibers in the focal plane due to optimization at different wavelengths (Margala et al., 2016).

The use of mock spectra with improved modeling of metal absorbers (Sect. 3), including both Lymetal and metalmetal correlations.

Modeling of the distortions of the correlation function due to quasarcontinuum fitting (Sect. 4). This allows us to construct a physical model of the correlation function over the range .

Modeling of spurious correlations introduced by the pipeline (Sect. 5), calculation of their effect on the correlation function, and searches for unidentified spurious correlations using the CarbonIV (CIV) forest, nm.

Fits of the data (Sect. 6) that marginalize over the contributions to the correlation function of metals, unidentified highcolumndensity systems, and fluctuations of ionizing UV flux.
None of these improvements induce significant changes in the derived values of and ; the differences (Sect. 8) with those of Delubac et al. (2015) are consistent with statistical fluctuations induced by the increased sample size.
This paper is organized as follows. Section 2 describes the DR12 data used in this analysis. Section 3 gives a brief description of the mock spectra used to test the analysis procedure, with a more detailed description being found in Bautista et al. (2015). Section 4 presents our method of estimating the fluxtransmission field, its correlation function, and the associated covariance matrix. Section 5 studies spurious correlations induced by the pipeline. In Section 6 we fit the mocks and data to derive the BAO peak position parameters, and . Section 7 investigates possible systematic errors in the measurement. In Section 8 we compare our measured peak position with the predictions of CDM models and derive constraints on the associated cosmological parameters. Section 9 presents a brief summary.
2 The BOSS quasar sample and data reduction
The quasar sample was gathered over a fiveyear period by the SDSSIII Collaboration (Eisenstein et al., 2011; Gunn et al., 1998, 2006; Smee et al., 2013). We use the data from the twelfth Data Release (DR12) of SDSS as presented in Alam et al. (2015). The associated quasar catalog is described in Pâris et al. (2016). Most of the quasar spectra were obtained by the Baryon Oscillation Spectroscopic Survey, BOSS (Dawson et al., 2013), but DR12 also includes six months of data from the SEQUELS ^{1}^{1}1http://www.sdss.org/dr12/algorithms/ancillary/boss/sequels/ program. The DR12 celestial footprint covering is displayed in Fig. 1. An example of a quasar spectrum in the Lyforest region is shown in Fig. 2. The redshift distribution of measurement pairs in the forest is shown in Fig. 3.
The quasar target selection used in BOSS, summarized in Ross et al. (2012), combines different targeting methods described in Yèche et al. (2010), Kirkpatrick et al. (2011), and Bovy et al. (2011). The selection algorithms use SDSS photometry and, when available, data from the GALEX survey (Martin et al., 2005) in the UV; the UKIDSS survey (Lawrence et al., 2007) in the NIR, and the FIRST survey (Becker et al., 1995) in the radio.
The DR12 data were processed using a new software package that differs from the standard DR12 SDSSIII pipeline (Bolton et al., 2012) and which has become the standard pipeline for SDSS DR13 (SDSS Collaboration et al., 2016). For each object, both pipelines provide a flux calibrated spectrum, , errors, and an object classification (galaxy, quasar, star). A model spectrum is fit to providing a redshift estimate. For this study, we use the “coadded” spectra constructed from typically four exposures of 15 minutes resampled at wavelength pixels of width (). For the small number of quasars with repeated observations, the coadded spectra can include exposures widely separated in time.
An important difference with respect to the DR12 pipeline is that pixels on the CCD image are combined to give a flux with pixelweights determined only by the CCD readout noise. While this method is suboptimal because it ignores photoelectron Poisson noise, compared to the DR12 method it yields an unbiased estimate of since the weights do not depend on the observed CCD counts which are needed to estimate Poisson noise. A more detailed description of the changes to the extraction pipeline is given in Appendix A.
The ratio of observed flux to model flux, averaged over all spectra, is an important diagnostic of pipeline systematic errors. Figure 4 shows the average ratio, , as a function of observed wavelength for quasar spectra on the red side of Ly emission. Since model imperfections for individual quasar spectra are averaged over in this figure (because of the range of quasar redshifts), the deviations from unity reflect imperfections in the spectrograph flux calibration. The figure reveals percentlevel deviations that are mostly due to imperfect modeling of photospectroscopic standard stars. The calcium H and K lines from interstellar absorption are also visible. In the analysis to be presented here, the are divided by to correct on average for these artifacts. This procedure is effective only with the fluxes from the DR13 pipeline since the nonlinearities present in the DR12 pipeline make the correction fluxdependent. We show in Section 5 that after this global correction, remaining calibration artifacts (due to their timedependence) are sufficiently small to have a negligible effect on the measurement of the correlation function.
The spectra of all quasar targets were visually inspected (Pâris et al., 2012, 2014, 2016) to correct for misidentifications, to flag broad absorption lines (BALs), and to establish the definitive quasar redshift. Damped Ly troughs (DLAs) were visually flagged, but also identified and characterized automatically (Noterdaeme et al., 2012). The visual inspection of DR12 confirmed 297,301 quasars, of which 181,719 are in the redshift range appropriate for this study, 2.1 3.5. We discarded quasars with visually identified BALs (to avoid the necessity of modeling their profiles in the forest) leaving 160,868 quasars. A further cut requiring a minimum number of unmasked forest pixels (50 “analysis pixels”; see below) yielded a sample of 157,922 quasars. Finally, 139 spectra failed the continuumfitting procedure (Sect. 4.1), leaving 157,783 spectra compared to 137,562 in the Delubac et al. (2015) investigation.
For the measurement of the flux transmission, we use the restframe wavelength interval
(1) 
As illustrated in Fig. 2, this range is bracketed by the emission lines nm and nm. This region was chosen as the maximum range that avoids the large pixel variances on the slopes of the two lines due to quasartoquasar diversity of lineemission strength and profile. The absorber redshift, , is required to lie in the range . The lower limit is set by the requirement that the observed wavelength be greater than 360 nm, below which the system throughput is less than 10% of its peak value. The upper limit is produced by the maximum quasar redshift of 3.5, beyond which the BOSS surface density of quasars is not sufficient to be useful for this study. The weighted distribution of redshifts of absorber pairs near the BAO peak position is shown in Fig. 3. The distribution has a mean of .
For the determination of the correlation function, we use analysis pixels that are the inversevarianceweighted flux average over three adjacent pipeline pixels. Throughout the rest of this paper, “pixel” refers to analysis pixels unless otherwise stated. The width of these pixels is corresponding at to an observedwavelength width and a comoving radial distance of . The total sample of 157,783 spectra thus provides measurements of Ly absorption over an effective volume of .
Mocks  Planck  
(TT+lowP)  
0.1323  0.1426  
0.1096  0.1197  
0.0227  0.02222  
0  0.0006  
0.7  0.6731  
3  3  
0.795  0.829  
0.97  0.9655  
0.27  0.3147  
[Mpc]  149.7  147.33 
(104.80 )  (99.17 )  
38.63  39.15  
8.744  8.612 
3 Mock quasar spectra
In order to test the analysis procedure and investigate statistical and possible systematic errors, we created 100 sets of mock spectra that reproduce the essential physical and instrumental characteristics of the BOSS spectra. The basic method for the production of the mocks with Ly absorption is described in FontRibera et al. (2012). Except for the implementation of absorption due to metals and high columndensity systems (HCDs, i.e., damped Lysystems and Lymanlimit systems), the mocks used in this study are identical to the DR11 mocks (Bautista et al., 2015) that were used in Delubac et al. (2015). The DR11 mocks therefore cover less solid angle than the DR12 data.
For each set of spectra, the background quasars were assigned the angular positions and redshifts of the DR11 quasars. The unabsorbed spectra (continua) of the quasars were generated using the Principal Component Analysis eigenspectra of Suzuki et al. (2005). The amplitudes for each eigenspectrum were randomly drawn from Gaussian distributions with a dispersion equal to that of the corresponding eigenvalues in Table 1 of Suzuki (2006). The overall normalization was chosen by fitting the mock spectrum to the corresponding observed spectrum.
The Ly absorption field was generated using the technique described in FontRibera et al. (2012) in which a Gaussian random field, , is first defined at the positions of the observed forest pixels. Flux transmissions were defined by the nonlinear transformation
The two functions, a(z) and b(z) are chosen so that the resulting mean transmission and variance are near those observed in the data. The correlations of the field are chosen so that the correlations of follow the linear correlations of the “mock” cosmology of Table 1 modified by nonlinear effects (McDonald, 2003). For pixels randomly distributed in space, this procedure would involve inverting a matrix. To reduce the problem to a manageable size, use was made of the fact that the forest pixels are nearly parallel, allowing a separate treatment of radial and transverse coordinates.
In the final step, the spectra were modified to include the effects of the BOSS spectrograph point spread function (PSF), readout noise, photon noise, and flux systematic errors.
For each of the 100 mock data sets, four types of spectra were produced and analyzed. The first type consists simply of Ly fluxtransmission fraction, , modified for the wavelength resolution but without multiplication by a quasar continuum spectrum, . Analysis of this mock type allowed us to study the recovery of the BAO peak position under the most favorable conditions. With the introduction of the quasar continuum, the second type consists of more realistic spectra, . Analysis of this type tests our ability to fit the quasar continuum and to model the resulting distortion of the correlation function.
The final two types of spectra add to absorption that is due to HCDs and to metals. Following FontRibera & MiraldaEscudé (2012), HCDs are placed at randomly chosen pixels where the optical depth was above a chosen threshold. The neutralhydrogen column densities were drawn randomly with , assuming an intrinsic powerlaw distribution corrected for selfshielding and normalized to match the observations of Prochaska et al. (2005). Because the HCDs are placed in redshift space, the resulting HCDs have correlations that trace the underlying density field but with a redshiftspace distortion parameter that is not necessarily equal to that of physical HCDs.
Absorption due to metal transitions was also simulated by adding absorption in proportion to the Ly absorption at the same redshift. The important transitions can be seen in the “one dimensional” correlation function, for fluxes at two wavelengths within the same forest, as shown in Fig. 5. In this figure, the peaks correspond to absorption by two different transitions by material at the same physical position but a different wavelength. The most prominent features are due to pairs comprising Ly and one metal transition, but there are also features due to metalmetal pairs. More detailed information can be obtained by stacking spectra around strong absorbers in the forest (Pieri et al., 2010, 2014). Table 2 lists the transitions included in the mock spectra.
We have tested two procedures for adding metal absorption to the primary Ly absorption. The methods are defined by the assumed relation between Ly absorption and metal absorption at the same physical position. For each wavelength, , with Ly absorption , an appropriate metal absorption for each transition, , is chosen. A simple procedure would be to choose with a proportionality constant chosen to reproduce the Lymetal features in the observed . However, this procedure does not produce significant features corresponding to metalmetal correlations. To do this it is necessary to add, for a randomly chosen small fraction of wavelengths, a much larger metal absorption.
The first procedure (hereafter Met1) aims to match the observed lineofsight correlation function , in particular the amplitudes of the peaks associated to Lymetal and metalmetal correlations. We first transform the pure mock Ly flux into “optical depth” and define the metal absorption of transition as . A quadratic term in is added for a fraction of the strong Ly absorbers to simulate strong metal absorption. The parameters , the quadratic terms and rates of strong absorption are set to match the observed amplitudes in the data .
The second procedure (hereafter Met2) aims to match the observed stack of high signaltonoise ratio Ly absorbers, while maintaining consistency with the in the resulting mock . Following Pieri et al. (2014), stacks were produced using the DR12 Ly forest sample. As a function of the Ly transmission, we measured amplitudes for the metal absorption features. The obtained “flux decrements” caused by metals were treated as target average flux decrements. Metals are implemented in Met2 mocks as a mix of weak and strong metal absorption, generating a mock consistent with the observed and consistent with the absorption frame measurements of Pieri et al. (2014).
The metal absorption added by these methods is due to the presence of metals at the same physical position as HI. Because our mocks provide the HI density only for redshifts , we cannot generate absorption at lower redshifts that nevertheless appears in the Ly forest because of transition wavelengths much greater than that of Ly. An important example is CIV doublet ( nm) where the absorption at an observed wavelength of 400 nm is due to material at . Fortunately this absorption has little effect on the correlation function, as we will see in Sect. 6.
The statistical properties of metal absorption in the mocks are determined by the underlying density field. However, the analysis procedure interprets absorption at a given wavelength as absorption due to the Ly transition. Because of this behavior, the metal contribution to the measured correlation function is shifted and deformed in space. In particular, the large correlation due to HI and metals at the same physical position is seen at with where is the relative wavelength separation of the metal feature with respect to Ly. This leads to a large correlation at this (and ), so if the amplitude is significant, it can be measured. Table 2 lists the apparent corresponding to vanishing Lymetal separation.
transition  [] 

SiIII(120.7 nm)  21 
SiIIa(119.0 nm)  64 
SiIIb(119.3 nm)  56 
SiIIc(126.0 nm)  111 
4 Measurement of fluxtransmission field and its correlation function
In this Section we describe the measurement of the correlation function of the transmitted flux fraction:
(2) 
Here, is the observed flux density for quasar at observed wavelength , is the unabsorbed flux density (the socalled “continuum”) and is the mean transmitted fraction at the absorber redshift, . Measurement of the fluxtransmission field requires estimates of the product for each quasar. An example is shown in Fig. 2. The estimation procedures, described in this Section, differ slightly from those of our previous studies (Busca et al., 2013; Delubac et al., 2015). One important modification is that we now calculate the distortion of the correlation function due to continuumfitting (Sect. 4.3).
4.1 Estimation of
As in our previous investigations, we assume the quasar continuum, , is the product of a universal function of the restframe wavelength, , and a linear function of , included to account for quasar spectral diversity:
(3) 
with being normalized so that its integral over the forest is equal to unity. The and are determined by maximizing the likelihood function given by
(4) 
Here is the probability to observe a flux for a given continuum found by convolving the intrinsic probability, , with the observational resolution assumed to be Gaussian:
(5) 
Here, is the variance due to readout noise and photon statistics.
A continuumdetermination method is defined by the assumed form of . For this work, it is taken to be the lognormal model of absorption used to generate the mock data, corresponding to “method 2” of our previous studies (Delubac et al., 2015). As a check, we also use “method 1”, which is equivalent to using a narrow Gaussian for , thereby producing only the product for each forest.
In practice, we maximize the likelihood iteratively by assuming a to determine the . The mean absorption is then calculated by an appropriately weighted average of (for fixed ) after which is recalculated as an average of (for fixed ). The procedure is stopped after ten iterations, at which point a stability of is reached for and . Figure 2 shows a spectrum with its and .
The function to be used in Eq. 2 to calculate is the mean of and can thus differ slightly from used in the model to estimate the . The use of the mean of ensures that the mean of is zero at each redshift (or wavelength).
Figure 6 displays the calculated , both for the DR13 pipeline used here and the DR12 pipeline used previously. The use of the new pipeline removes most of the artifacts in the old analysis. We emphasize, however, that the derived is dependent on the assumed form of and should therefore not be considered as a measurement of the mean absorption. For example, the flattening of for () suggests that we have slightly underestimated at these redshifts. Since, by construction, the mean vanishes at each redshift, this implies that our procedure makes a compensating overestimate of the . Since it is the product that determines , the measured correlation function is therefore not strongly affected.
Those forests with identified DLAs are given a special treatment. All pixels where the absorption due to the DLA is higher than 20% are not used. The absorption in the wings is corrected using a Voigt profile following the procedure of Noterdaeme et al. (2012).
We denote as the estimate of using the relation (2). Because forest data is used to fit for and the measured is not equal to the original . We can identify two effects. First, the use of is equivalent to the transformation
(6) 
where the overbar refers to the average over forests at fixed . Second, the fitting of with the data biases toward zero the mean and mean . To simplify this effect and facilitate its correction we explicitly subtract the mean and first moment for each forest
(7) 
where the overbars refer to averages within a given forest using the weights defined in the next Section. The accompanying distortion of the correlation function is nonnegligible, as we will demonstrate in Sect. 6.2 with the mock spectra.
The transformation (7) has two interesting effects on the measured fluxtransmission field. Most importantly, it makes it simple to calculate the distortion of the correlation function (Sect. 4.3) and thus simplifies the relation between the underlying physical model and the measured correlation function. Second, it nearly eliminates the difference between the correlations functions calculated with the two continuum fitting methods used in Delubac et al. (2015) with the r.m.s. difference in the two being 0.056 of the r.m.s. uncertainty per bin.
4.2 Estimation of the correlation functions
For the estimator of the flux autocorrelation function, we adopt a simple weighted sum of products of the :
(8) 
where the are weights (see below) and each (or ) indexes a measurement on a quasar at wavelength . The sum over is understood to run over all pairs of pixels within a bin in the space of pixel separations, . We exclude pairs of pixels from the same quasar to avoid the correlated errors in and arising from the estimate of for the spectrum of the quasar. The bins are defined by a range of width of the components perpendicular and parallel to the line of sight, and . We use 50 bins in each component, spanning the range from to ; the total number of bins used for evaluating the correlation function is therefore 2500. Separations in observational pixel coordinates (RA,Dec,) are transformed to in units of by assuming that absorption is due to the Ly transition and using the cosmological parameters from Table 1 (Planck cosmology for the data and the mock cosmology for the mocks).
As described in Delubac et al. (2015), the weights, , are chosen so as to account for both Poisson noise in the flux measurement and for the intrinsic fluctuations in due to cosmological largescale structure. The weights are set to zero for pixels flagged by the pipeline as having problems due, for example to sky emission lines or cosmic rays. To reduce the pipeline systematics discussed in Sect. 5, we also do not use pairs of pixels that have nearly the same wavelength () and that were taken on the same exposures.
4.3 The distortion matrix
The transformations (6) and (7) mix pixels so that the correlation is equal to the original plus a linear combination of the correlations of other pixelpairs . This approach means that the measured correlation function is a “distorted” version of the true correlation function . Since the transformations (6) and (7) are linear, the relation between measured and true correlation functions is given by a distortion matrix :
(9) 
where and refer to bins in pixel separation space. Writing produces
(10) 
where . We ignore the small effect of transformation (6), in which case unless the pixels and are in the same forest:
(11) 
where the two sums over include only pixels in the same forest as that of and and the overbars refer to averages in that forest. The matrix thus depends only on the geometry and weights of the survey. We will see its effect on the mock correlation function in Sect. 6.2 (Fig. 11).
Previous analyses (Busca et al., 2013; Slosar et al., 2013; Delubac et al., 2015; Blomqvist et al., 2015) have dealt with the distortions introduced by continuumfitting in different ways. Busca et al. (2013) and Delubac et al. (2015) model it as an additive power law in and with 12 free parameters. Blomqvist et al. (2015) assumed that continuumfitting reduces the observed amplitude of longwavelength modes in the direction parallel to the line of sight. They then model it as a multiplicative function of that tends to zero at large scales (with , where is the typical length of a forest) and to one at small scales (). They tune the shape of the function using simulated data to ultimately reduce the number of free parameters to one.
In our approach we do not introduce free parameters to account for the effects of continuumfitting. Instead, we follow the assumption, first proposed by Slosar et al. (2013), that at each line of sight the continuumfit delta field differs from the true delta field by a linear function in . Slosar et al. (2013) then deweight these “linear modes” in their covariance matrix. Alternatively, we use the transformation from to to remove the “linear modes” from the model via the distortion matrix.
4.4 The covariance matrix
The covariance matrix associated with is:
(12) 
where . Following Delubac et al. (2015), we evaluate by dividing the BOSS footprint into subsamples and measuring and in each subsample . Neglecting the small correlations between subsamples, and replacing the fourpoint function by the measured product of correlations in subsamples, the covariance is given by
(13) 
where the are the sums of weights in the subsample . As in Delubac et al. (2015), the SDSS plates define the subsamples.
As a check of the subsampling method, the sum (12) can also be estimated by neglecting interforest correlations, in which case the fourpoint function vanishes unless the four pixels are drawn from just two spectra:
(14) 
The sum can then be estimated from a random sample of forest pairs. Because neighboring forests are nearly parallel, the sum necessarily gives unless .
Finally, for the analysis of the mock data, the covariance matrix can also by calculated from the mocktomock variations of :
(15) 
where the overbar refers to averages over the set of 100 mocks. Given the approximations used in the calculation of via eqns. 13 or 14, it is of great importance that the mocktomock variations confirm the accuracy of the other two methods when applied to mock data. This comparison is shown in Fig. 7.
The element matrix, , has a relatively simple structure. By far the most important elements are the diagonal elements which are, to good approximation, inversely proportional to the number of pixel pairs used in the calculation of the correlation function; the number of pairs is roughly proportional to :
(16) 
The variance is about twice as large as what one would calculate assuming all (analysis) pixels used to calculate are independent. This decrease in the effective number of pixels is due to the physical and instrumental correlations between neighboring pixels in a given forest (eqn. 14).
The offdiagonal elements of the covariance matrix also have a simple structure. As previously noted, the covariance is mostly due to the twoforest part of the fourpoint function which, because neighboring forests are nearly parallel, only contribute to the covariance matrix elements with . This behavior is illustrated in Fig. 7, which displays the mean values of the correlation matrix elements as a function of for the smallest values of .
The statistical precision of the subsampling calculation is for individual elements of the correlation matrix. Figure 7 reveals that only correlations with and are greater than the statistical precision and therefore sufficiently large for individual matrix elements to be measured accurately by subsampling. As in Delubac et al. (2015), we therefore average the subsampling correlation matrix over and use the resulting covariance matrix that is a function of .
5 Correlations introduced by the optics and data pipeline
Spurious correlations in are introduced by the telescope and spectrometer optics, and by the pipeline reductions of the data. These correlations are superimposed on the physical that we wish to measure; in this Section we estimate the various contributions.
5.1 Optical crosstalk
At the optical level, correlations are introduced by signals from one object “scattering” into the spectra of other objects, that is through optical crosstalk. There is negligible crosstalk introduced in the telescope focal plane by photons from one object entering into the fiber of another object. However, there can be measurable crosstalk between neighboring fibers downstream of the wavelength dispersion where they are focused on the CCD. This contamination arises through imperfect modeling of the pointspread function when transforming the twodimensional CCD image into a series of onedimensional spectra, . We have measured the crosstalk as a function of fiber separation by fitting the signal in sky fibers to a sky model and a weighted sum of the spectra of the objects (quasars or galaxies) in neighboring fibers. Consistent with the results of Croft et al. (2016), we find the crosstalk for the neighboring fibers is and for fibers separated by two or three rows respectively, and is consistent with zero for larger separations.
The crosstalk directly introduces correlations between pixels at the same wavelength but these pixel pairs are, at any rate, not used in the analysis. The crosstalk introduced between pixels and of two quasars is proportional to the product of the crosstalk amplitude and to . We have verified that these correlations are insignificant compared to the measured . This fact is in part due to the fiberassignment strategy which avoided placing two quasar candidates on neighboring fibers.
5.2 Pipelineinduced correlations
The pipeline treatment of the spectrometer data transforms flatfield corrected CCD counts, , to fluxes, . This process requires subtracting a sky contribution, , and multiplying by a calibration vector, , that corrects for the wavelengthdependent throughput of the system:
(17) 
Both and are determined from spectra taken simultaneously with the data: spectra of spectrophotometric standard stars for , and spectra of “sky fibers” (fibers pointing to empty sky regions) for . The spectra from the 1000 fibers of a given exposure are treated independently for the two 500fiber spectrographs because of possible differential variations of instrumental properties such as throughput, PSF, and scattered light. (See Smee et al. (2013) for the relevant details of the spectrometer construction.) Hereafter, the collection of fibers of a given plate assigned to one of the two spectrographs will be referred to as a “halfplate.” Inaccuracies in the determinations of and will lead directly to correlated inaccuracies in the of a given halfplate.
Errors in the sky subtraction, , are simple to model because they are mostly due to wellunderstood Poisson fluctuations of photoelectron counts in the sky fibers per halfplate. The pipeline interpolates these measurements in the focal plane to produce the sky background to be subtracted from a given quasar. The limited number of skyfibers results in correlations between and for and since the sky subtractions where calculated using the same noisy skyfibers. These correlations for generate small correlations for through continuum fitting, as discussed in Sect. 4.2. Because the correlations are primarily for they contribute a spurious for pairs of pixels on the same halfplate.
Errors on the calibration vector are also Poissonian in that they are due to the fluctuations in the imperfect modeling of calibration stars per halfplate. The mean over all spectra of the imperfections are visible in the mean “transmission” of the unabsorbed parts of quasar spectra shown in Fig. 4. As mentioned in Section 2, the mean imperfections are removed by dividing forest spectra by from Fig. 4. However, fluctuations from exposure to exposure remain because of the spectral variations of the small number of calibration stars used. The statistical properties of these variations have been estimated by comparing the measured with random subsamples of stars and, independently, by comparing the on different halfplates for the same exposure. Unlike the correlations due to skysubtraction, which are confined to , the errors in the calibration vectors are strongly correlated over the characteristic wavelength range of stellar spectral features, nm.
We constructed a model for pipelineinduced correlations that uses the measured statistical properties of the and . The contribution of the sky noise was estimated using a fit of random realizations of the signal in the sky fibers (using a degree 2 polynomial fit as a function of the fiber number). The realizations were based on the statistical uncertainty estimation from the pipeline (which we have found to be accurate at the 5% level, see Fig. 19, bottom panel). A normalization factor was applied to account for the average difference of calibration between the sky and target fibers (a correction). Performing a fit allowed us to capture accurately the resulting correlations as a function of fiber separation.
The contribution of the calibration was estimated using the fact that because the flux calibration of the two spectrographs is determined independently, the difference of those two calibration solutions for each exposure provides us with an estimate of their statistical fluctuation (up to a normalization factor ) while retaining the correlation of those fluctuations as a function of wavelength. Those differences for all DR12 plates was computed. Their average per observation run was subtracted in order to account for the actual difference of throughput of the two spectrographs and their evolution during the course of the survey. This data set provided us with a library of calibration uncertainties that was used to calculate calibrationinduced correlations between spectra.
This model was used to calculate the expected correlation between pixels in the Ly forest of one quasar and pixels in the CIV forest ( nm) of another quasar. The physical correlation of the two forests is due mostly to the autocorrelation of the weak CIV absorption in the two forests, but any pipelineinduced correlations should be present at full strength. Since this crosscorrelation is designed to isolate pipelineinduced effects, we denote this correlation function as where are the wavelengths in the two forests and is the angular separation. We also distinguish between and correlations between flux pairs on the same halfplate, , and correlations between flux pairs on different halfplates, . While these correlations are naturally given as a function of wavelength and angular separations, for convenience, the will be given as functions of pseudoseparations calculated using the Ly restframe wavelength to define redshifts in both spectra. With this approach, corresponds to absorption at the same observed wavelength.
This correlation for samehalfplate pairs (measured using the techniques of Sect. 4.2) is shown with the red points in Fig. 8 for the first bin (). Superimposed on the data is the prediction of the model of the pipelineinduced correlations. For the first bin, the correlation is dominated by those induced by to the skysubtraction model. There is good agreement between this simple model and the observed correlations.
The pipelineinduced correlations that we have considered do not contribute to correlations between pixels observed on different halfplates. The differenthalfplates correlation functions for the bin is shown by the blue points in Fig. 8. The correlations for different halfplates (blue points) are clearly far less than those for same halfplates (red points). For , () for the nocorrelation hypothesis.
For the bins, the model predicts much smaller correlations. In particular, the skysubtraction model noise induces nonzero correlations only because of the continuum fit which distorts the original correlation function as described in Section 4.2. This behavior is illustrated in Fig. 9 showing in four ranges of . Because of the low level of absorption fluctuations redward of Ly emission, the variance of the points is a factor smaller than the corresponding variance for the forestforest correlation function.
The model for pipelineinduced correlations assumes that the skysubtraction errors are entirely due to Poisson statistics of photoelectron counts, neglecting possible wavelength and positiondependent systematic misestimates of the sky flux. While the changes applied to the pipeline (Appendix A) considerably reduced the systematic sky residuals, significant residuals remain on bright sky lines due to an imperfect PSF model, PSF variations and displacement of spectral traces (due to changes of temperature and variations of the gravity load of the spectrographs which are at the Cassegrain focus of the telescope). We tested the effect of those sky residuals by computing the correlation function of fake Ly forests consisting of the residual signal in the nearest sky fiber in the CCD divided by the quasar continuum model (we made sure to use only once each sky fiber in this process to avoid introducing an artificial correlation). The measured correlation signal could be entirely explained by the Poissonian sky model noise described above. We hence conclude that the systematic sky residuals induce a negligible contamination to the Ly correlation function.
This analysis presented above evaluates the contamination due to any additive signal to the Ly forest that leads to a correlated signal in the CCD. It includes the systematic sky residuals, their fluctuations from plate to plate, but also the potential effect of scattered light in the spectrograph.
We have used our model of pipelineinduced correlations to calculate their effect on the determination of the BAO peak position. Since the model contains no scale near the BAO scale, it is not surprising that it predicts no measurable influence, as reported in Table 7 of Sect. 7. Furthermore, to facilitate the comparison of the measured correlation function with the physical model developed in the next section, we do not use pairs of pixels on the same exposure that would contribute to the bins of .
parameters  best fit  best fit 

Ly  
Si3  
Si2a  
Si2b  
Si2c  
CIV  
HCDs  
6 Fits for the BAO peak position
To determine the position of the BAO peak in the transverse and radial directions, we fit the measured to functions that describe the underlying largescalestructure correlations. These correlations are primarily due to Ly absorption in the intergalactic medium (IGM), but we also include absorption by metals in the IGM and neutral hydrogen in highcolumndensity systems (HCDs). The physical correlations are corrected for distortions that are introduced by the procedure for determining the quasar continuum (Sect. 4.3). The fitting routine we use gives results consistent with those found using the publicly available baofit^{2}^{2}2http://darkmatter.ps.uci.edu/baofit/ package (Kirkby et al., 2013; Blomqvist et al., 2015) modified to include the effects described in this Section. The parameters of the fits are described below and in Table 3. The bestfit model correlation function is shown in Fig. 10.
6.1 The model of the correlation function
The general form for the model correlation function of the bin is a distorted sum of the cosmological, or “physical” correlation function, , and a slowly varying function, , used to test for systematics:
(18) 
Here, is the distortion matrix (eqn. 10) that models the effects of continuum fitting.
The physical component of the model is dominated by the autocorrelation due to Ly absorption. It is assumed to be a biased version of the total matter autocorrelation of the appropriate flatCDM model (Table 1) modified to free the position of the BAO peak (Kirkby et al., 2013):
(19) 
where the BAO peakposition parameters to be fit are
(20) 
and where the subscript “fid” refers to the fiducial cosmological model from Table 1 used to transform angle differences and redshift differences to . The nominal correlation function, , is derived from its Fourier transform
(21) 
where is the (quasi) linear power spectrum decomposed into a smooth component and a peak component corrected for nonlinear broadening of the BAO peak:
(22) 
The smooth component is derived from the CAMBcalculated linear power spectrum, , via the sideband technique (Kirkby et al., 2013) and . The correction for nonlinear broadening of the BAO peak is parameterized by . The nominal values used are and (Kirkby et al., 2013).
In equation 21, the bias, , is assumed to have a redshift dependence with (so that ), and is assumed redshift independent. The function corrects for nonlinear effects at large due to the isotropic enhancement of power due to nonlinear growth, the isotropic suppression of power due to gas pressure, and the suppression of power due to lineofsight nonlinear peculiar velocity and thermal broadening. We use the form given by equation 21 and Table 1 of McDonald (2003). The forms proposed by ArinyoiPrats et al. (2015) produce nearly identical results for the range of used in this study.
The last term in equation 21, , accounts for the binning in space which effectively averages the correlation function over a bin. (The large width of these bins renders unnecessary a term that accounts for the spectrometer resolution.) If the distribution of observed were uniform in a bin, would be the Fourier transform of the bin. The distribution is approximately uniform in the radial direction, which implies where (the bin width). In the perpendicular direction, the distribution is approximately proportional to so