Abstract
We provide an updated recommendation for the usage of sets of parton distribution functions (PDFs) and the assessment of PDF and PDF+ uncertainties suitable for applications at the LHC Run II. We review developments since the previous PDF4LHC recommendation, and discuss and compare the new generation of PDFs, which include substantial information from experimental data from the Run I of the LHC. We then propose a new prescription for the combination of a suitable subset of the available PDF sets, which is presented in terms of a single combined PDF set. We finally discuss tools which allow for the delivery of this combined set in terms of optimized sets of Hessian eigenvectors or Monte Carlo replicas, and their usage, and provide some examples of their application to LHC phenomenology. This paper is dedicated to the memory of Guido Altarelli (1941–2015), whose seminal work made possible the quantitative study of PDFs.
Communicated by Alan D Martin
1. Introduction
In this first section we introduce the general context for the updated PDF4LHC 2015 recommendations, and describe the layout of this document. Users whose main interest is the application of the PDF4LHC15 recommendations to their specific analysis can move directly to section 6.
1.1. Parton distributions at the LHC
The accurate determination of the parton distribution functions (PDFs) of the proton is crucial for precision predictions at the large hadron collider (LHC) [1–3]. Almost all cross-sections of interest are now available at next-to-leading order (NLO), a rapidly increasing number also at next-to-next-to-leading order (NNLO), and even one, inclusive Higgs production in gluon fusion, at NNNLO [4]. The resulting improvements in theoretical uncertainties from the inclusion of higher order matrix elements demand a correspondingly improved control of PDF uncertainties.
A number of groups have recently produced updates of their PDFs fits [5–11] that have been compared to LHC data. Even for PDFs based on similar data sets, there are still variations in both the ensuing central values and uncertainties. This suggests that use of the PDFs from one group alone might underestimate the true uncertainties, and a combination of individual PDF sets is required for a robust uncertainty estimate in LHC cross-sections.
1.2. The PDF4LHC working group and the 2010 recommendations
The PDF4LHC working group has been tasked with:
- (1)performing benchmark studies of PDFs and of predictions at the LHC, and
- (2)making recommendations for a standard method of estimating PDF and PDF + uncertainties at the LHC through a combination of the results from different individual groups.
This mandate has led to several benchmarking papers [12, 13] and to the 2010 PDF4LHC recommendation [14] which has undergone several intermediate updates, with the last version available (along with a summary of PDF4LHC activities) from the PDF4LHC working group website: http://hep.ucl.ac.uk/pdf4lhc/.
In 2010 the PDF4LHC working group carried out an exercise to which all PDF groups were invited to participate [12]. Benchmark comparisons were made at NLO for LHC cross-section predictions at 7 TeV using MCFM [15] as a common framework using carefully prescribed input files. The benchmark processes included W/Z total cross-sections and rapidity distributions, cross-sections and Higgs boson production through gg fusion for masses of 120, 180 and 240 GeV. The PDFs used in this comparison included ABKM/ABM09 [16], CTEQ6.6/CT10 [17, 18], GJR08 [19, 20], HERAPDF1.0 [21], MSTW2008 [22] and NNPDF2.0 [23]. The results were summarized in a PDF4LHC report [12] and in the LHC Higgs cross section working group (HXSWG) Yellow Reports 1 [24] and 2 [25]. In this study, each group used their native value of along with its corresponding uncertainty.
This 2010 PDF4LHC prescription [14] for the estimation of combined PDF+ uncertainties was based on PDFs from the three fitting groups performing a global analysis with a variable flavor number scheme, namely CTEQ, MSTW and NNPDF. Here, 'global analysis' is meant to signify that the widest available set of data from a variety of experiments and processes was used, including deep-inelastic scattering, gauge boson Drell–Yan (DY) production, and inclusive jet production at hadron colliders.
The recommendation at NLO was to use the envelope provided by the central values and PDF+ errors from CTEQ6.6, MSTW2008 and NNPDF2.0, using each group's prescription for combining the PDF with the uncertainties [26–28], and with the central value given by the midpoint of the envelope. By definition, the extent of the envelope is determined by the extreme PDFs at the upper and lower edges. A drawback of this procedure was that it assigned a higher weight to these outlier PDF sets than what would be statistically correct when all the individual input sets have equal prior likelihood.
In addition, as each PDF uses its own native value of and its own PDF + uncertainties about that central value, the resultant envelope inflates the impact of the uncertainty. This very conservative prescription was adopted partly because of ill-understood disagreements between the PDF sets entering the combination (with other sets differing even more), and it was considered suitable specifically for the search of the Higgs boson at the LHC. Furthermore, it was suggested that in case of updates of the various PDFs set by the respective groups, the most recent sets should always be used.
At the time of this first recommendation, only MSTW2008 (of the three global fitting groups) had produced PDFs at NNLO. Since PDF errors are determined primarily by the experimental errors in the data sets, as well as by the methodology used in the corresponding PDF extraction (such as tolerance), it was known that PDF uncertainties were relatively unchanged when going from NLO to NNLO. The NNLO recommendation was thus to use MSTW2008 NNLO as the central result, and to take the same percentage error on that NNLO prediction as was found at NLO using the uncertainty prescription described above.
1.3. Intermediate updates
A follow-up benchmarking study in 2012 was carried out with the NNLO versions of the most up-to-date PDF sets [13]. By that time, CTEQ6.6 had been replaced by CT10 [18, 29] and NNPDF2.0 by NNPDF2.3 [30], both at NNLO. In addition, HERAPDF1.0 had been replaced by HERAPDF1.5 [31], and ABM09 by ABM11 [32]. In that study, a common value of was used by all PDFs, with a variation between 0.117 and 0.119 to account for the uncertainty on
Data from the LHC Run I was then already available, and detailed comparisons were made to data for inclusive jet and electroweak boson production from ATLAS [33, 34], CMS [35] and LHCb [36]. Much better agreement was observed between the updated versions of the three PDF sets used in the first study, CT10, MSTW2008, and NNPDF2.3, for example for the quark–antiquark PDF luminosity in the mass region of the W/Z bosons, see figure 1. However some disagreements remained: figure 1 also shows the gluon–gluon PDF luminosity for MX values in the region of the Higgs boson mass. In this case, the resulting envelope was more than twice the size of the uncertainty band for any of the individual PDFs.
It is interesting to observe that the uncertainty bands for CT10, MSTW2008 and NNPDF2.3 are all reasonably similar, as shown in figure 2, even though the methodologies used to determine them differ. The uncertainties for ABM11 are smaller, despite using a more limited data set, and also including sources of uncertainties due to and heavy-quark masses; we will come back to this issue in section 2.3 below. The uncertainties for HERAPDF1.5 were substantially larger, as expected from the more limited data set used.
Download figure:
Standard image High-resolution imageBased on the 2012 study, the 2010 PDF4LHC recommendation was updated in 2013. First, it was recommended that the most up-to-date versions of the PDF sets from the three groups included in the previous recommendation be used, namely CT10, MSTW2008 and NNPDF2.3. Furthermore, as all sets now had both NLO and NNLO sets, it was recommended that the same procedure should be used both at NLO and NNLO. Finally, a somewhat simpler way of combining the PDF and uncertainties was suggested. Namely, the central value of was fixed for all PDFs to be obtained by rounding off the then-current PDG world average 0.1184 ± 0.0007 [37] (and near the preferred value of each group anyway). An uncertainty range for was taken to be ±0.002 at the 90% confidence level (CL) around the central value of 0.118. This corresponds to a 68% CL uncertainty of ±0.0012, somewhat more conservative than the PDG estimate.
The total PDF uncertainty was then determined for each group by adding in quadrature the PDF uncertainty and uncertainty, with the latter determined as the difference in the results found using the best-fit PDFs for the upper and lower values of the 68% CL range. Indeed, it can be shown [27] that addition of PDF and uncertainties in quadrature automatically accounts for the correlation between and PDF uncertainties, assuming gaussianity and linear error propagation. As in the 2010 recommendation, it was then suggested to determine the PDF+ uncertainty for each of the three groups, at the upper and lower value of the chosen range and finally take the envelope of the results.
1.4. Scope of this document
The scope of the current document is to provide an update of the PDF4LHC recommendation suitable for its use at the LHC Run II. Similar to previous recommendations, the new recommendation will be based on a discussion and comparison of existing PDF sets. Specifically, we will argue that, based on more recent results, an envelope procedure is no longer necessary and a purely statistical combination is appropriate. In addition, we will specify criteria for the inclusion of PDF sets in this combination. We will then discuss dedicated tools which allow for a streamlined construction and delivery of this statistical combination, and finally discuss the main feature of a combined PDF4LHC PDF set constructed in this way.
The outline of the paper is as follows. In section 2 we review developments in PDF determination since the original PDF4LHC recommendation, and in section 3 we compare the most recent PDF sets from all groups. In section 4 we motivate the new prescription for the combination of PDF sets and enumerate criteria for the inclusion of PDF sets in this combination. In section 5 we discuss the practical implementation of the 2015 PDF4LHC prescription, based on the combination of the CT14, MMHT14 and NNPDF3.0 PDF sets. First we introduce the Monte Carlo (MC) statistical combination, and we then present various methods suitable for the production and delivery of a manageable reduced combined PDF set, based on either MC or Hessian methodologies. We also present the resulting PDF4LHC15 combined sets, both at the level of individual PDFs and parton luminosities, and use them to compute several representative LHC cross-sections, specifically comparing the various delivery forms.
For the majority of the users of the new PDF4LHC15 recommendations, the only section which is essential is section 6. This section presents a self-contained summary with the general guidelines for the usage of the PDF4LHC15 combined sets, including the formulae for the computations of PDF and PDF+ uncertainties, and the corresponding citation policy. Finally, conclusions are drawn and directions for future developments are presented in section 7.
2. Recent developments in PDF determination
Since the 2010 PDF4LHC recommendation document [14] there has been very considerable progress in the determination of PDFs. This includes both the usage of an increasingly large and diverse set of data, exploiting new measurements from the LHC, Tevatron and HERA experiments, and various methodological developments, which have been incorporated in various new PDF sets released by different groups. In this section we will give a brief summary of these developments.
2.1. Intermediate PDF updates
Some of the updates in PDF sets were fully expected at the time of the last recommendation document, and, as mentioned in section 1, have been already implemented in updated recommendations, available at the PDF4LHC website [38].
Specifically, at the time of the previous recommendation the most updated set from NNPDF was NNPDF2.0 [23], obtained using the zero-mass variable-flavor-number scheme, which is known to miss important mass-dependent terms for scales near the quark mass and to induce inaccuracies in the PDFs. However, soon after the recommendation this set was updated to NNPDF2.1 [39] which used the FONLL general-mass variable flavor number scheme (GM-VFN scheme) described in [40], and this automatically improved general agreement with CT and MSTW PDFs. These PDFs were later also made available at NNLO [41]. Subsequently NNPDF updated their PDFs to the NNPDF2.3 set [30], the main change being the inclusion of early LHC data on rapidity-dependent vector boson production and asymmetries and on inclusive jet production. In addition, the theoretical treatment of charged-current structure functions in neutrino DIS was improved in NNPDF2.3, leading to a somewhat enhanced strangeness around Even though changes from NNPDF2.1 to NNPDF2.3 are moderate, the new data and improved fitting methodology reduced uncertainties a little.
In 2013 the CT10 NNLO PDFs were released, specifically using the extension to NNLO [42] of the S-ACOT-χ heavy flavor scheme used by this group. This enabled an intermediate improved prescription based on the envelope of CT10, MSTW08 and NNPDF2.3 at NLO and NNLO, as discussed in section 1.
Coming to MSTW08, it was soon clear that despite good predictions for most LHC observables, the MSTW2008 PDFs did not describe low rapidity lepton asymmetry data, which is sensitive to small-x () valence quarks. This was rectified in an intermediate update, the MMSTWW (or sometimes MSTWCPdeut) PDFs [43], based on the same data sets as MSTW2008, but with an extended PDF parametrization based on Chebyshev polynomials in rather than just powers of and more flexible deuteron correction (given as a function with four free parameters that are fit to data). This included a study of the number of terms in the parametrization of quarks and gluons required to reproduce functions to an accuracy of small fractions of a percent: seven parameters were found to be sufficient. These PDFs agree well with LHC W asymmetry data, even though the data itself is not actually included in the fit.
There have also been intermediate updates from groups not included in the 2010 PDF4LHC combination. PDF sets from the HERAPDF group are based on fitting to HERA data only. While the HERA structure function data is undoubtedly the most constraining single data set available, particularly for the gluon and the total quark singlet distribution at moderately and small x, it provides fewer constraints in some kinematic regions (such as the gluon at large x) and some PDF combinations, and specifically it does not allow a separation of the down and strange distribution. At the time of the last recommendation, the most recent available PDF set was HERAPDF1.0, which was obtained from a fit to the HERA Run-I combined total cross-section data. This set was in good broad agreement with CT, MSTW and NNPDF sets, but with larger uncertainties for many PDF regions and a much softer high-x gluon and harder high-x sea quarks. Since then, there have been a number of updates. HERAPDF1.5 [31] included some preliminary data on inclusive cross sections from HERA-II running. This reduced uncertainties slightly, and also led to a harder high-x gluon and sea-quark distributions at NNLO, though not as much at NLO. These differences between NLO and NNLO can be partly explained by the fact that the NNLO sets appeared some time after the NLO ones, and also that the NNLO HERAPDF1.5 set was based on a more flexible PDF parametrization that its NLO counterpart.
There are also two other PDF sets that are based on smaller total sets of data than the CT, MMHT and NNPDF fits. These PDFs differ from all those discussed so far in that they adopt a fixed-flavor-number (FFN) scheme for the treatment of heavy quarks: i.e., they do not introduce PDFs for charm and bottom quarks and antiquarks. The most recent sets from these groups at the time of the previous recommendation were ABKM09 [16] and (G)JR09 [19, 44]. The group responsible for the former, now the ABM collaboration, has since produced two updated sets. The first one was the ABM11 PDFs [32], which included the combined HERA Run-I cross-section data for the first time, and also some H1 data at different beam energies in addition to the fixed target DIS and DY data already used in the ABKM analysis. The PDFs were mostly unchanged, but with a slightly larger small-x gluon. As in the ABKM analysis a rather low value of i.e. at NNLO, was obtained.
In [45], the impact on the ABM11 analysis of fitting different jet datasets from D0, CDF and the 2010 ATLAS data was studied. This was done both at NLO and at NLO* (including the effects of threshold resummation) in jet production. As compared to the default fit, once the Tevatron jet data sets were included, the fitted value of was found to vary by an amount depending on the particular data set and jet algorithm. These jet measurements have not been included in the posterior ABM PDF releases.
All these PDF fitting groups have presented major updates recently: MMHT14 [10], CT14 [6] NNPDF3.0 [11], ABM12 [5] and HERAPDF2.0 [9], which will be discussed in more detail in section 2.3 below.
2.2. New experimental measurements
One obvious reason for updates of the various PDF sets is the availability of new measurements from a variety of collider and fixed-target experiments. Recently, an up-to-date overview of all PDF-sensitive measurements at LHC Run-I has been presented in the 'PDF4LHC report on PDFs and LHC data' [1]. We refer to this document for an extensive discussion (including full references) and here we only provide a very brief summary.
First of all, HERA has provided combined H1 and ZEUS data on the charm structure functions, in addition to the already available combined inclusive Run-I structure function data. Run-II ZEUS and H1 data have also been published, and a legacy HERA combination of all inclusive structure function measurements from Run-I and Run-II has been presented very recently [9]; it is only included in the HERAPDF2.0 set. Also, Tevatron collaborations are still releasing data on lepton asymmetry and top-pair cross-sections and differential distributions.
A wide variety of data sets which help constrain the parton distributions has been made available at the LHC from the ATLAS, CMS and LHCb collaborations. This includes data sets on rapidity- and mass-dependent and production, inclusive jet cross-sections, inclusive top-pair production and charm quark production, differential top pair production and dijet cross sections.
Many of these measurements have been included in recent global PDF sets. Therefore, for all PDF groups the fitted dataset is now considerably wider than that at the time of the 2010 PDF4LHC recommendation.
2.3. Current PDF sets and methodological improvements
We now discuss in turn all the presently available PDF sets. These differ from previous releases not only because of the inclusion of new data, but also because of various methodological and theoretical improvements, which we also review.
We first consider the most updated PDF sets from the three groups represented in the original recommendation. An important observation is that the most recent sets from these groups, CT14, MMHT2014 and NNPDF3.0 sets are all made available with full PDF uncertainty information available at the common value of both at NLO and NNLO. This is to be contrasted to the situation at the time of the previous recommendation, when full information of PDF uncertainties was only available in each set for fixed (different) values of spanning a range of approximately We now discuss each set in turn.
The CT14 PDF sets [6] have been made recently available at NLO, NNLO, and also at LO. These sets include a variety of LHC data sets as well as the most recent D0 data on electron charge asymmetry [46] and combined H1+ZEUS data on charm production [47]. The PDFs also use an updated parametrization based on Bernstein polynomials, which reduces parameter correlation. The PDF sets contain 28 pairs of eigenvectors. LHC inclusive jet data are included at NLO also in the NNLO fit, despite the lack of a full NNLO calculation, in the expectation that the NNLO corrections are relatively small for single-inclusive cross-sections. The main change in the PDFs as compared to CT10 is a softer high-x gluon, a smaller strange quark (partially due to correction of the charged current DIS cross-section code) and the details of the flavor decomposition, e.g. and the high-x valence quarks, due both to the parametrization choices and new data. In the CT14 publication [6] a number of additional comparisons were made to data not included in the fit, such as W + charm and from the LHC, and the predictions are found to be in good agreement. One should also mention that the CT14 PDF uncertainties are provided as 90% CL intervals, that need to be rescaled by a factor 1.642 to compare with other PDF sets, for which uncertainties are provided as 68% CL intervals.
We now consider PDFs from the MSTW group, currently renamed MMHT due to a change in personnel. The MMHT14 PDFs [10] incorporate the improved parametrization and deuteron corrections in the MMSTWW study [48], and also a change in the GM-VFN scheme to the 'optimal' scheme in [49], and a change in the branching fraction used in the determination of the strange quark from data. Charged-current structure functions were also updated in the MMHT14 analysis, in particular by including the NLO gluon coefficient functions. The updated analysis includes new data: the combined HERA-I inclusive structure function data [21] (which post-dates the MSTW2008 PDFs) and charm structure function data, updated Tevatron lepton asymmetry data, vector boson and inclusive jet data from the LHC (though LHC jet data is not included at NNLO), and top pair cross-section data from the Tevatron and LHC. No PDFs change dramatically in comparison to MSTW2008, with the most significant changes being the shift in the small-x valence quarks already observed in the MMSTWW study, a slight increase in the central value of the strange quark to help the fit to LHC data, and a much expanded uncertainty on the strange distribution due to the inclusion of a conservative uncertainty on The MMHT14 PDFs provide a much better description to the LHC W lepton asymmetry data [34, 50] than its predecessor MSTW08, thanks to the various improvements already implemented in the MMSTWW (also known as MSTW08CPdeut) intermediate release. Deuteron nuclear corrections are fitted to the data, and their uncertainty propagates into the total PDF uncertainty.
The MMHT14 PDFs are made available with 25 eigenvector pairs for and 0.120 at NLO and 0.118 at NNLO, as well as at LO (for ). However, is also determined by the NLO and NNLO fits and values of and 0.1172 respectively are found, or 0.1195 and 0.1178 if the world average of is included as a data point. These are in good agreement with the PDG world average without DIS data of [107]. A dedicated study of the uncertainties in the determination of in the MMHT14 analysis has been presented in [51].
The NNPDF3.0 PDF sets [11] are the recent major update within the NNPDF framework. In comparison to NNPDF2.3, NNPDF3.0 includes HERA inclusive structure function Run-II data from H1 and ZEUS (before their combination), more recent ATLAS, CMS and LHCb data on gauge boson production and inclusive jets, and charm and top quark pair production. A subset of jet data was included at NNLO using an approximate NNLO treatment, based on a study [52] of the region where the threshold approximation [53] is reliable by comparing with the exact gg channel calculation [54]. Heavy quarks are treated using the FONLL-B scheme (FONLL-A was used in NNPDF2.1 and NNPDF2.3), which includes mass corrections to an extra order in thereby achieving a better description of low-Q2 data on the charm structure function. The compatibility of the NNPDF3.0 NLO set with the LHCb forward charm production cross-sections has been demonstrated in [55], where it was also shown that these data could be useful in reducing the uncertainty on the small-x gluon.
The NNPDF3.0 fitting procedure has been tuned by means of a closure test, namely, by generating pseudo-data based on an assumed underlying set of PDFs. One verifies in this case that the output of the fitting procedure is consistent with the a priori known answer. As a byproduct, one can investigate directly the origin of PDF uncertainties, and specifically how much of it is due to the uncertainty in the data, how much it is due to the need to interpolate between data points, and how much is due to the fact that there is an infinite number of functions which produce equally good fits to a given finite number of data-points. The minimization has been optimized based on the closure test, specifically by choice of a more efficient genetic algorithm, and the choice of optimal stopping point based on cross-validation and the search for the absolute minimum of the validation (look-back fitting). The NNPDF3.0 PDFs display moderate changes in comparison to NNPDF2.3: specifically somewhat smaller uncertainties and a noticeable change in the gluon–gluon luminosity.
The HERAPDF2.0 PDFs has also become recently available [9]. HERAPDF2.0 is the only PDF set to include the full legacy Run-I and Run-II combined HERA structure function data, based on runs at beam energies of and 460 GeV. This PDF set has considerably reduced uncertainties on PDFs compared to HERAPDF1.0. In particular there is a much improved constraint on flavor decomposition at moderate and high x due to the difference between neutral current and cross-sections, and due to much more precise charged current data, which provides a genuine constraint on the dV distribution. The running at different energies gives sensitivity to which provides some new information on the gluon. The HERAPDF2.0 sets are made available also with a default of and with 14 eigenvector sets along with further variations to cover uncertainties due to model assumptions and changes in the form of the parametrization; for CT, MMHT and NNPDF this is not necessary because the input form of the PDFs is flexible enough. Sets including HERA jet and charm data have also been made available. The HERAPDF2.0 PDFs agree fairly well in general with the CT, MMHT and NNPDF sets, but there are some important differences in central values, particularly for high-x quarks. Overall, the uncertainties on HERAPDF2.0 PDFs are markedly smaller than for previous versions, but there are still some regions where the PDFs have significantly larger uncertainties than those in the global fits.
Let us also mention that the three GM-VFN schemes used by these four PDF sets (S-ACOT-χ, FONLL and RT/RTopt) exhibit a remarkable convergence, especially at NNLO [56]. All these four sets fit the HERA charm cross-sections with comparable quality, and residual differences between the three GM-VFN schemes translate into rather small differences in the resulting PDFs [57].
The most recent update for the ABM collaboration, ABM12 [5], now includes the HERA combined charm cross-section data, an extension of the HERA inclusive data to higher Q2 than previously used, and vector boson production data from ATLAS, CMS and LHCb. The heavy flavor contributions to structure function data are now calculated using the renormalization scheme [58] rather than the pole mass scheme. The main change compared to ABM11 PDFs is in the details of the decomposition into up and down quarks and antiquarks, this being affected by the LHC vector boson data. The ABM12 PDF sets are determined together with whose value comes out to be rather lower than the PDG average, namely at NNLO. Top quark pair production data from the LHC is investigated, but not included in the default PDFs. Its inclusion tend to raise the high-x gluon and a little; the precise details depend on the precise value of the top quark mass (and mass renormalization scheme) used. The ABM12 PDFs are available with 28 eigenvector pairs for the default and at NNLO only. Within the same framework there has been a recent specific investigation of the strange sea [59] in the light of LHC data and NOMAD and CHORUS fixed-target data, but this is not accompanied by a PDF set release. More recently, an update of the ABM12 with additional data from gauge boson production at the Tevatron and the LHC has been presented [60], though this study is not accompanied by the release of a new PDF set.
There has also been an update of the (G)JR PDFs: JR14 [8]. The default sets from this group make the assumption that PDFs must be valence-like at low GeV2. This analysis extends and updates the fixed-target and collider DIS data used, in particular including CCFR and NuTeV dimuon data as a constraint on strange quarks, and includes both HERA and Tevatron jet data at NLO, but does not include any LHC data. The strong coupling is also determined along with the PDF, yielding a value at NNLO, and a large small-x gluon distribution. There are some significant changes compared to the JR09 NNLO PDFs. These are mainly at high x values, where there is an increase in down and strange quarks and antiquarks, and a decrease in the gluon. At low x values, the features are largely determined by the assumption of a valence-like input, and are largely unchanged. A determination without the valence-like assumption (and thus a more flexible parametrization) is also performed, leading to PDFs and a best fit value of more in line with other groups, though it is not adopted as a default.
In addition, there has been a number of more specific PDF studies. The CJ12 PDFs [7] emphasize the description of the high-x region, with particular care paid to higher-twist corrections and the modeling the deuteron corrections to nucleon PDFs, thereby allowing for the inclusion of high-x and low Q2 data which is often cut from other fits, most notably Jefferson Lab data. Several analyzes have been focused on the possible intrinsic charm component of the proton. Specifically, [61, 62] have presented investigations of the evidence for, and limits on, intrinsic charm in the proton. Work in progress towards the determination of intrinsic charm is also ongoing within NNPDF, and the required modifications of the FONLL GM-VFN scheme have been presented in [63]. The determination of mass of the charm quark from PDF fits has been discussed in [64, 65]. PDFs including QED corrections, which were presented for the first time in [66] by the MRST group, have been updated in [67]; a determination of the photon PDF from the data has been presented for the first time in [68], with the corresponding study in the CT14 framework found in [69].
It is also important to mention that a number of PDF-related studies based on the open-source HERAfitter framework [70] have also become available. From the HERAfitter developers team, two studies have been presented: PDFs with correlations between different perturbative orders [71] and the impact on PDFs of the Tevatron legacy DY measurements [72]. Within the LHC experiments, various studies of the constraints of new measurements on PDFs have been performed using HERAfitter, including the ATLAS analysis [73] that lead to a strangeness determination and the CMS studies constraining the gluon and from inclusive jet production [74, 75]. The impact of the LHCb forward charm production data on the small-x gluon has been explored by the PROSA collaboration in [76]. Finally, HERAfitter has also been used to quantify the impact on PDFs of data from a future large hadron electron collider [77].
2.4. Origin of the differences between PDFs
There have been many developments in understanding the differences between the PDF sets obtained by different groups. This understanding inevitably leads to an improvement of the best procedures for combining these different PDFs. In the benchmark document [12], on which the first PDF4LHC recommendation was based, the differences between the PDFs and the consequent differences in predictions for LHC cross-sections were discussed; subsequent benchmarking exercises were published in [32, 78]; and then in [13] including comparisons with published LHC measurements. See also [79] for an earlier benchmark exercise. Here we concentrate on three topics on which some progress has been made recently: the impact of the choice of heavy-quark scheme; the origin of the remaining differences between global PDF fits; and the origin of the differences between the sizes of PDF uncertainties. Graphical comparisons of the PDFs illustrating their present level of agreement are presented in section 3.
2.4.1. Dependence on the heavy-quark scheme
The results of various benchmark studies [13, 32, 78] illustrate how the softer high-x gluon and smaller present in PDF sets like ABM12 and JR09 leads to a smaller prediction for top pair cross sections and Higgs bosons as compared to the global PDF sets. In both cases, the magnitude of the difference depends on whether the default is obtained directly from the fit or if it is treated as an external input. In these sets also the quark PDFs seem to lead to slightly larger vector boson production cross-sections as compared to the global fits. On the other hand the CT, MSTW/MMHT, and NNPDF, and HERAPDF PDFs at NNLO are in generally good agreement, though HERAPDF has associated rather larger uncertainties from the reduced dataset.
In the previous recommendation it was stated that the systematic feature of a lower high-x gluon and lower value of in some fits was due to the omission of Tevatron jet data, which constrains the high-x gluon. The understanding of this issue has improved. While it is true that Tevatron and LHC jet data directly constrain the gluon distribution, and that PDF sets with a particularly small high-x gluon distribution and/or small value of do not provide the best fit to these data (see [45, 80] for discussions on this issue with similar results, though not the same conclusions), it is unclear whether their omission automatically leads to a small high-x gluon distribution or small The results of including the jet data in the fit also show some dependence on the treatment of correlated systematic errors (additive or multiplicative) assumed, though these effects are way too small [11, 29] to explain the aforementioned differences.
The behavior of the gluon is constrained directly by Tevatron/LHC jet production at high x, and indirectly by scaling violations in DIS data at very small and high x, with the latter influencing the gluon at moderate x via the momentum sum rule. The gluon PDF is strongly correlated with at high x and anti-correlated at low x. Due to this feature of correlation between the gluon and there is no automatic strong pull to lower or small high-x gluon in the fits that omit hadron collider data [80, 81]. It appears that the DIS and DY data provide only a weak constraint on and the gluon, if both are left free, a conclusion also evident in those HERAPDF1.6 studies that added HERA jet data to the DIS data and more strongly constrain
An analysis in the ABM framework [82] suggested that this issue is related to whether one makes a conservative cut on low-Q2 structure function data, or includes it and includes higher-twist corrections as well as higher-order QCD corrections to in fits of the NMC cross-section data. However, their findings were not corroborated by CTEQ, MSTW and NNPDF groups [80, 83–85], who did not observe clear-cut sensitivity of the PDFs to the details of the NMC treatment either at fixed [29, 83], or varying [29, 80]. Also, no significant variation in PDFs was seen when including a fixed higher twist in the NNPDF2.3 analysis in [84]. In [85], when the higher-twist contribution is fit within the standard MSTW framework, there was fairly small impact on the PDFs, and less on
However, different schemes for heavy-quark treatment, notably the usage of the fixed-flavor scheme as opposed to GM-VFN scheme, may explain the most pronounced disagreements observed in the gluon distribution between ABM/JR and the global sets. When a fit was performed to DIS data using the three-flavor FFN scheme at NLO using the MSTW and NNPDF frameworks [49, 84], the gluon became softer at high x and larger at small x; the light quarks became larger at small x (a feature previously noted to some extent in [86]); the value of decreased noticeably. This was confirmed in more detail in [85], using an approximate NNLO FFN scheme. At NNLO the decrease in was more marked that at NLO, and these conclusions did not change after the DY data and higher-twist corrections were added. It was also observed that the quality of the fit to the DIS data was clearly worse when using the FFN scheme, as opposed to the GM-VFN scheme. At a fixed the same conclusions about the PDFs, more detailed confirmation of a worse fit were found, while it was also observed that the deterioration is mainly in high-Q2 DIS data [84].
In [85] a theoretical investigation of the difference between the speed of evolution of at high Q2 was performed, and it was shown that, for the region there is slow convergence of the terms included at finite order in the FFNS to the fully resummed result in a GM-VFNS, leading to the worse fit to HERA data for this range of x at high Q2. It was also shown that this slower evolution is partially compensated by an increase of the gluon in this region of x, which feeds down to lower x with evolution. This, however, also results in a smaller gluon at high x, so that the fit to the fixed-target data then requires a lower value of This feedback between and the gluon is somewhat enhanced if higher twist contributions are included, as the freedom in these provides more flexibility to change the PDFs and without worsening the fit quality.
The general conclusion is thus that smaller values of (particularly at NNLO), softer high-x gluons, and slightly enhanced light quarks at high x are found if a FFNS, rather than GM-VFNS, is adopted in the PDF fit. In an extremely conservative approach, these differences might be viewed as a 'theoretical uncertainty', but the markedly worse fit quality for the FFNS motivates instead that a GM-VFNS should be adopted, especially for calculating quantities sensitive to high-Q2 PDFs.
2.4.2. Differences within the global PDF fits
While the above discussion sheds light on the differences between the sets based on a FFN scheme and those which adopt a GM-VFN scheme, the fact remains that CT, HERAPDF, MSTW and NNPDF also show some smaller differences, even though they all adopt a GM-VFN scheme. As mentioned in the introduction, for the previous families of each group, NNPDF2.3, CT10 and MSTW08, these differences were especially noticeable for the gluon PDF and led to a spread of predictions for Higgs production via gluon fusion, producing an envelope uncertainty that is more than twice the size of individual PDF uncertainties.
An attempt to understand the origin of these moderate differences was made in [57]. In this study, all groups produced fits to the HERA combined Run-I inclusive cross-section data only. Furthermore, all structure function calculation codes from the groups were benchmarked using a toy set of PDFs. The agreement was good in general, with all deviations in neutral current processes being due only to the differing choices in GM-VFNS choice, which were much smaller at NNLO than NLO due to the convergence of such schemes, and differing treatments of electroweak corrections at high Q2. A difference in charged current cross-sections was noted, and corrected, but it did not affect the PDFs extracted in any significant manner. Fits to the data were performed adopting a common choice of kinematic cuts, definition (i.e. treatment of correlated uncertainties) and charm mass, all of which differ to some extent in the default fits.
This study found that the PDFs and predictions for Higgs cross-sections at NNLO still followed the same pattern: in particular, for CT remained smaller than the MSTW result, which in turn was smaller than the NNPDF prediction, as in the global fits, even though all predictions were compatible within the large uncertainties of a HERA-only fit. An investigation of variations of the heavy flavor scheme was also performed, but this turned out to have a comparatively small impact. Some evidence of parametrization dependence between CT and MSTW was noted (and new parametrization forms were adopted by CT), but it was not conclusive.
Concurrent to this study with DIS data, NLO theoretical calculations for fitting collider jet data, based on programs MEKS [87] and NLOJET++ [88, 89], and their fast interfaces APPLgrid and FastNLO [90–93], were validated [13, 94, 95], providing greater confidence in the NLO theoretical codes and their fast interfaces, and the fits to the jet data.
Therefore, the details of the reasons for the disagreement at, or around, the one-σ level, between the global fits for the gluon luminosity in the region of the Higgs boson mass were still only partially understood. This was a situation that conservatively could only be handled by taking an envelope of results, which amounts to saying that one or more of the results are affected by an unknown source of systematics (or bias).
This state of affairs has changed completely with the release of the most recent CT14, MMHT14 and NNPDF3.0 PDF sets. Indeed, due to increased data constraints and methodological improvements in each of the fitting procedures, these turn out to be in better agreement for most PDFs in a wide range of x, as will be illustrated in section 3.
2.4.3. Differences in the size of the PDF uncertainties
In general the sizes of the PDF uncertainties between the three global sets, NNPDF, CT and MMHT, turn out to be in rather reasonable agreement. On the other hand, the sizes of the uncertainties for PDF sets based on a HERA-only dataset, such as HERAPDF, are be expected to be larger, and in some cases much larger, than those of the global fits. This is indeed what happens for the HERAPDF2.0 PDFs, despite including the legacy HERA combination data [9], since deep-inelastic scattering cannot constrain a number of quark flavor combinations and the medium and large-x gluon. The much larger PDF uncertainties in HERA-only fits as compared to global fits have also been verified in the NNPDF framework. The NNPDF3.0 HERA-only fit [11] indeed shows substantially larger uncertainties for most PDF combinations (with the exception of the quark singlet and the small-x gluon) than the corresponding global analysis.
Whereas HERAPDF2.0 includes the legacy HERA combination data [9], which are not included in any other fit, it has been shown [96, 97] that the impact of these data on a global fit that already included the HERA-I data was moderate. In addition, it has been checked [97] that their impact on a global fit like NNPDF3.0 which already included both the HERA-I combination and the individual HERA-II measurements is very small. Therefore, the availability of the legacy HERA inclusive dataset does not modify the general anticipation that HERA-only fits (or more general, DIS-only fits) will have larger PDF errors than the global fits. However, it is worth emphasizing that the uncertainties of HERAPDF2.0 are in many cases comparable (for the u quark) or even smaller (for the gluon) to the global fits, as will be illustrated in the comparisons of section 3.
The PDF uncertainties in ABM12 are systematically smaller, and often much smaller, that those of the global fits. Therefore the PDF uncertainty estimates of the three global fits, on one hand, and those of ABM12, on the other, do not appear to be consistent. The most likely explanation for the disparate sizes of the uncertainties is their varied statistical interpretations adopted by the PDF analysis groups. In the global sets, they are determined according to complex procedures, and include more sources of uncertainties than the purely experimental uncertainty obtained from the criterion, for example, arising from partial inconsistencies of the input datasets, using a variety of methods [10, 11, 98–101].
3. Comparisons of PDF sets
Let us now present a comparison of PDFs and parton luminosities, using the most up-to-date versions of the PDF sets. All the PDFs and luminosity comparison plots shown in this document have been obtained with the APFEL on-line PDF plotting tool [102, 103]. We will show comparisons for the ABM12, CT14, HERAPDF2.0, MMHT2014 and NNPDF3.0 NNLO PDF sets. For HERAPDF2.0, the uncertainty shown includes experimental, model, and parametrization uncertainties. For ABM12, we use the Nf = 5 set and the native value of since sets for different values of the strong coupling are not available.
3.1. Parton distributions
We begin by comparing the gluon and the up quark PDFs, with the corresponding one-sigma uncertainty bands. These are shown in figure 3 for Q = 100 GeV, normalized to the central value of the CT14 distribution. Similar plots are shown for the d and quarks in figure 4. For each flavor, the left plots compare the three global sets among them, while the right plots compare CT14 with the two sets based on reduced datasets, HERAPDF2.0 and ABM12.
Download figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageIn general, the gluon distributions for CT14, MMHT2014 and NNPDF3.0 agree well, but not perfectly, both in central value and in uncertainty, except at very low x and at high x. This level of agreement is reassuring, given the importance of the gluon distribution in the dominant Higgs boson production mechanism. Although each group uses similar, but not identical, data sets, the fitting procedures and tolerances are not the same, and for this reason, exact agreement is not expected. As compared to the global fits, the HERAPDF2.0 gluon PDF is somewhat larger in the x range from 10−4 to 10−2 and is substantially smaller for Similar considerations apply to ABM12, where their differences as compared to the global sets are due only in part to the difference value, with the bulk of the differences due to the use of a FFNS, as discussed in section 2.4 above.
The overlap for the up quark distribution is not as good as for the gluon distribution, so an envelope of CT14, MMHT2014 and NNPDF3.0 would be somewhat broader than the individual predictions. The HERAPDF2.0 and ABM12 up quark distribution are higher than the global sets, especially in the x range from 10−2 to 0.5, while HERAPDF2.0 agrees well with CT14 for smaller values of x. The CT14, MMHT2014 and NNPDF3.0 down quark distributions, see figure 4, are in reasonable agreement with each other over most of the x range, but have slightly different behaviors at high x. The HERAPDF2.0 down quark has a different shape, though its uncertainty bands partially overlap with those of the others. ABM12 agrees with the global fits except around where it is systematically larger.
Similar conclusions as those drawn from the comparison of u quark PDFs between the three global fits. apply to the quark distributions. Note the blow-up of PDF uncertainties at large x due to the reduced experimental information available in this region. The HERAPDF2.0 quark distribution is generally higher than in the global fits for by a substantial amount (around 10%). Here ABM12 agrees reasonably with CT14, except for where is substantially smaller.
3.2. PDF luminosities
It is also instructive to examine the parton–parton luminosities [104], which are more closely related to the predictions for LHC cross-sections. The gluon–gluon and quark–antiquark luminosities, as a function of the invariant mass of the final state MX, for a center-of-mass energy of 13 TeV are shown in figure 5, where we compare, on the left, the three global fits, NNPDF3.0, CT14 and MMHT14, and, on the right, CT14 with the fits based on reduced datasets, HERAPDF2.0 and ABM12, using for the latter exactly the same settings as in the PDF comparison plots. All results are shown normalized to the central value of CT14, as before. The corresponding comparison for the quark–quark and gluon–quark PDF luminosities is then shown in figure 6.
Download figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageThe luminosity uncertainty ranges tend to blow-up at low invariant masses ( GeV) and high masses ( GeV for gg, TeV for and TeV for qq), that is, in the regions that are relatively unconstrained in current global PDF fits. The region of intermediate final-state invariant masses can be thought of as the domain for precision physics measurements, where the PDF luminosity uncertainties are less than 5% (at 68% CL). There is good agreement among the predictions of CT14, MMHT2014 and NNPDF3.0 for the gg PDF luminosities for this mass range, and in particular for the production of a Higgs boson at
This is an improvement over the situation with the previous generation of PDFs (CT10, MSTW2008 and NNPDF2.3), where, as mentioned previously, the total PDF luminosity uncertainty in the Higgs mass range was more than a factor of 2 larger the uncertainty for any individual PDFs, as a result of differences in central PDFs. Clearly, this leads to a corresponding improvement in agreement between predictions for the cross section computed using these PDF sets. Indeed, as compared to NNPDF2.3, the new NNPDF3.0 prediction decreases by about 2%, the CT14 prediction increases by 1.1%, compared to CT10 while the MMHT14 predictions, as compared to MSTW08, decrease by a small amount (roughly 0.5%). As discussed in section 2.3, it is difficult to pinpoint a single reason for this improvement in agreement, which most likely arises from a combination of methodological advancements and new experimental constraints in the global fits.
Still considering the gg luminosities, HERAPDF2.0 is in good agreement with the global fits for TeV, becoming rather softer above this mass, while ABM12 shows strong differences, with a much harder luminosity for and a much softer one above this MX value. For the luminosity, there is a reasonable agreement between the three global fits; note the blow-up of PDF uncertainties at small and large values of MX because of absence of data. HERAPDF2.0 best-fit luminosity for is above CT14 by about 5% in the region of intermediate MX, with better agreement at smaller and larger masses, but markedly less certain. ABM12 is consistent with the global fits except for 2 TeV when it becomes rather softer.
Turning to the quark–quark and gluon–quark PDF luminosities shown in figure 6, we see that for the global fits we get consistent results within uncertainties, though the agreement is not quite as good as for the gluon–gluon luminosity, especially in the region between 100 GeV and 1 TeV. The luminosities for HERAPDF2.0 and ABM12 in the qq case are harder than those of the global fits, by around 5% at low masses to up 15% in the TeV region, with important phenomenological implications. For the qg luminosity, there is good agreement for the global fits (similar to the gg luminosity) and slightly worse for the PDF sets based on reduced datasets.
4. Constructing the PDF4LHC15 combination
As discussed in section 3, we are now in the rather satisfactory position where differences between PDF sets are either better understood or much reduced. However, there is still the question of how best to combine PDF sets even if they are essentially compatible. In this section we motivate and present the updated PDF4LHC prescription for the evaluation of PDF uncertainties at the LHC Run II, discuss the general conditions that PDF sets should satisfy in order to be included in the combination and its future updates, and list the PDF sets which will enter the current prescription. We then describe how the construction of the combined sets based on the MC method.
4.1. Usage of PDF sets and their combinations
We would like first to state that there are three main cases in which PDFs are used in LHC applications:
- (1)Assessment of the total uncertainty on a cross section based on the available knowledge of PDFs, e.g., when computing the cross section for a process that has not been measured yet (such as supersymmetric particle production cross-sections), or for estimating acceptance corrections on a given observable. This is also the case of the measurements that aim to verify overall, but not detailed, consistency with standard model expectations, such as when comparing theory with Higgs measurements.
- (2)Assessment of the accuracy of the PDF sets themselves or of related standard model parameters, typically done by comparing theoretical predictions using individual PDF sets to the most precise data available.
- (3)Input to the MC event generators used to generate large MC samples for LHC data analysis.
In the second case, it is important to always use the PDF sets from the individual groups for predictions. This is especially true in comparisons that involve PDF-sensitive measurements, providing information about the PDFs for individual fits. As a rule of thumb, comparisons between QCD or electroweak calculations and unfolded data for standard model production processes should be done with individual PDF sets. Similar considerations hold for the third case, since MC event generators require to be carefully tuned to experimental data using as input specific PDF sets.
However, in the first case above a robust estimate of the PDF uncertainty must accommodate the fact that the individual PDF sets are not identical either in their central values or in their uncertainties. Consequently, an uncertainty based on the totality of available PDF sets must be estimated. Besides acceptance calculations, a contemporary example of this situation is the extraction of the Higgs couplings from LHC data, in which a robust estimate of the overall theoretical uncertainty is essential for probing the nature of the Higgs boson and for identifying possible deviations from the standard model expectations. In searches for new physics particles, such as supersymmetric partners or new heavy and bosons, estimates of the combined uncertainty are also necessary in order to derive robust exclusion limits, or to be able to claim a discovery.
For the applications of type 1, instead of using an envelope, the 2015 PDF4LHC recommendation proposes to take a statistical combination, to be described below, of those PDF sets that satisfy a set of compatibility requirements. While these requirements may evolve with time, for the current combination we select the individual PDF sets that satisfy the following properties:
- (1)The PDF sets to be combined should be based on a global dataset, including a large number of datasets of diverse types (deep-inelastic scattering, vector boson and jet production, ...) from fixed-target and colliders experiments (HERA, LHC, Tevatron).
- (2)Theoretical hard cross sections for DIS and hadron collider processes should be evaluated up to two QCD loops in, in a GM-VFN with up to active quark flavors15 . Evolution of and PDFs should be performed up to three loops, using public codes such as HOPPET [105] or QCDNUM [106], or a code benchmarked to these.
- (3)The central value of should be fixed at an agreed common value, consistent with the PDG world-average [107]. This value is currently chosen to be at both NLO and NNLO16 . For the computation of uncertainties, two additional PDF members corresponding to agreed upper and lower values of should also be provided. This uncertainty on is currently assumed to be again the same at NLO and NNLO. The input values of mc and mb should be compatible with their world-average values; either pole or masses are accepted.
- (4)All known experimental and procedural sources of uncertainty should be properly accounted for. Specifically, it is now recognized that the PDF uncertainty receives several contributions of comparable importance: the measurement uncertainty propagated from the experimental data, uncertainties associated with incompatibility of the fitted experiments, procedural uncertainties such as those related to the functional form of PDFs, the handling of systematic errors, etc. Sets entering the combination must account for these through suitable methods, such as separate estimates for additional model and parametrization components of the PDF uncertainty [9], tolerance [6, 10], or closure tests [11].
Following the needs of precision physics at the LHC, future updates of the PDF4LHC recommendations might be based on a different set of conditions. For instance, in addition to the above PDF sets, one might be required to provide fits in a range of mc and mb values, or to provide direct evidence of the ability to describe, reasonably accurately, a wide variety of different data types that constrain PDFs of different flavor and in different kinematic regions. In the future, with the progress in combination techniques, it might be possible to relax the first requirement, and also include in the combination PDF sets based on datasets of a different size by using suitable weighted averaging techniques—note that such a weighted averaging would have to account for the fact that different data affect different PDFs and kinematic regions, and thus the weights chosen cannot be the same for all PDF flavors.
The existing PDF sets which satisfy all of these requirements at present have been identified as CT14, MMHT2014 and NNPDF3.0; no other publicly available PDF sets currently appears to satisfy all conditions.
4.2. Statistical combination of PDF sets
Currently, two different representations of PDF uncertainties are being used [12]: the MC representation [108–110], in which the probability distribution of PDFs is given as an ensemble of replicas, whose mean and standard deviation provide respectively central value and uncertainty, and the Hessian representation [99], in which a central PDF is given, along with error sets, each of which corresponds to an eigenvector of the covariance matrix in parameter space.
In [101], a comparison of the Hessian and the MC methods was made within the MSTW framework, showing that they provide compatible representations of PDF uncertainties, and in particular lead to the same uncertainties when used to determine PDFs from known pseudo-data. Also, a way of obtaining an MC representation from starting Hessian representation was presented, and the accuracy of the Hessian to MC conversion was explicitly demonstrated. As will be discussed in the next section, recently two methods to perform the inverse conversion, namely transforming MC sets into a Hessian representation, have also been developed [111, 112].
These developments make possible a statistical combination of different PDF sets and their predictions, as originally outlined in [2]: if different PDF sets are assumed to be equally likely representations of an underlying PDF probability distribution, they can be combined by simply taking their unweighted average. This in turn can be done simply by generating equal numbers of MC replicas from each input PDF set, and then merging these replica sets of equal sizes. For the Hessian PDF sets such as CT or MMHT, the MC replicas are generated by sampling along each eigenvector direction, assuming a Gaussian distribution. Alternatively, a weighted average would correspond to taking different number of replicas from the various sets sets entering the combination. First examples of this application for LHC cross sections were given in [3]. A detailed overview of how the MC method has been used to construct the PDF4LHC 2015 combined sets will be presented in the next section.
Clearly, results obtained in this way are less conservative than those obtained from an envelope: they correspond to the assumption that different PDF determinations are statistically distributed instances of an underlying probability distribution, rather than instances of a probability distribution affected by unknown underlying sources of systematics. Such a combination method appears therefore to be adequate when results are compatible, or differences are understood, as is the case now.
5. Implementation and delivery of the PDF4LHC15 PDFs
In this section we discuss the technical construction of the new PDF4LHC prescription, presented in section 4, and the delivery of the combined PDFs based on it. Readers interested in the more practical question of how to use the PDF4LHC15 combination in their specific analysis can jump directly to section 6.
Unlike the previous PDF4LHC prescription, in which the task of combining the PDF uncertainties was left to the user, with the current PDF4LHC prescription, several pre-packaged PDF sets that already combine the uncertainties will be constructed and delivered. The combined sets are based on a statistical combination of the CT14, MMHT2014 and NNPDF3.0 PDF sets, as discussed in section 4.2. This combination leads to a 'prior' MC set with replicas, to be referred to as either MC900 or PDF4LHC15_prior in the following discussion. Such a large replica set would be unmanageable; however, various methods have been proposed recently to deal with it, which we collectively refer to as reduction methods. Usage of these methods will allow for a compact delivery of the combined PDF sets.
The idea of producing a unified PDF set by combining various individual sets was first suggested in [111], based on the idea of refitting a suitable functional form to a combined set of MC replicas, thereby leading to a representation of the starting MC probability distribution in terms of Hessian error sets in parameter space (META-PDFs). This then realizes the dual goal of producing a combination, and then reducing the number of PDF error members to a manageable size. More recently, other reduction methods were suggested, with the similar goal of turning the starting combined MC sample into a more manageable representation. The MC compression method [113] selects a subset of the original replica sample which reproduces its statistical features with minimal information loss, thereby keeping the MC representation, but with a smaller number of replicas (CMC-PDFs). The MC to Hessian conversion method [112] turns a set of MC replicas into Hessian error sets by representing the covariance matrix in the space of PDFs on a discrete set of points in x as a linear combination of the replicas, by means of a singular value decomposition (SVD) followed by principal component analysis (PCA) (MCH-PDFs).
In this section, we will first present the MC combination of the PDF sets and, in particular, determine the number of replicas that is necessary in order to achieve a faithful description of the combined set. We will then review each of the three reduction and delivery methods: CMC (Monte Carlo), META (Hessian) and MCH (Hessian), and, in each case, identify the size of the corresponding reduced PDF sets that optimizes the specific features of the method. These three sets will be adopted for delivery of the combined PDF4LHC15 PDF sets. Finally, we will compare and benchmark the PDF sets obtained with the three different reduction strategies, both at the level of parton distributions and of LHC observables.
5.1. The MC combination of PDF sets
The first step in the construction of an MC statistical combination [2, 3, 101, 111, 113, 114] is the transformation of the Hessian PDF sets into MC PDF sets using the Watt–Thorne method [101]. Once all sets have been turned into an MC representation, the combination is simply built by adding together an equal number of replicas from each set. The generation of the MC replicas is based on the symmetric Thorne–Watt formula, see equation (22) in [115], as implemented in the LHAPDF6 code [115]. This has been cross-checked with the independent code of the MP4LHC package, finding excellent agreement.
In figure 7 we show the comparison of the PDFs from the MC combination of CT14, MMHT14 and NNPDF3.0 for a different number of MC replicas, 900 and 1800, referred in the following as MC300, MC900 and MC1800. The error bands correspond to one standard deviation from the mean value. We see that, while between and some visible, albeit small, changes are observed, results stabilize if a yet larger number of replicas is used. We conclude that (i.e. 300 replicas from each of three individual sets) is an adequate number for the combination, and this is the value that we will adopt henceforth as a default for our prior. Therefore, the prior combined set PDF4LHC15_prior coincides with the MC900 set.
Download figure:
Standard image High-resolution imageIn figure 8 we compare the MC900 combined PDF set with the three individual PDF sets, CT14, MMHT14 and NNPDF3.0, at NNLO, for at GeV2. Results are normalized to the central value of MC900. Because of the good consistency of the PDF sets which enter the combination, the combined uncertainty is not much larger than that of individual PDF sets. A somewhat more significant increase observed in some cases, for instance, for the strange around due to the larger differences between the individual PDF sets. By construction, the uncertainty on the statistical combination remains always smaller than that of the envelope method.
Download figure:
Standard image High-resolution imageIn figure 9 we show the probability distribution for MC900 compared to that of the individual input PDF sets. Results are shown for the gluon at the up quark at the down antiquark at x = 0.2 and the strange PDF at x = 0.05. The histograms represent the probability per bin of each PDF value, quantified by the number of replicas that fall into that bin. For comparison, a Gaussian distribution with the same mean and variance as that of the MC900 histogram is also plotted. The combination appears to be generally Gaussian to good approximation, though some features of the combined distribution provide evidence of deviation from Gaussian behavior, for example a non-vanishing skewness in the and distributions.
Download figure:
Standard image High-resolution imageThe non-Gaussian features of the MC combination can also be observed for specific LHC cross-sections, in particular for extreme kinematic regions or for those processes that involve PDF combinations affected by large uncertainties. As an illustration, in figures 10 and 11 we show the probability distribution for MC900 using the Kernel density estimation method [113] for two specific observables. First of all, the W + charm differential cross-section in the lepton rapidity from the CMS measurement [116] in the range and then the forward DY process in dileptons from the LHCb measurement [36] in the range for These probability distributions are clearly non-Gaussian: double-hump structure for W + charm; significant skewness for LHCb forward DY. In the same plot, results obtained after applying two reduction techniques are also shown, to be discussed below.
Download figure:
Standard image High-resolution imageThe non-Gaussian features for W + charm production shown in figure 10 are related to the different assumptions on the parameterization of the strange PDF adopted by the various groups. For instance, the MMHT14 analysis allows the branching ratio to be determined from the fit, increasing the associated uncertainties in the strange PDF. Here the usefulness of the present combination method is clear, allowing to combine in a statistically consistent way the different assumptions used by each group (under the assumption that each of the various choices has equal prior likelihood—for the sets that enter the combination).
The discussion above refers to the combination of PDF sets for a common value of In addition, we need to provide combined sets for the values and in order to be able compute the combined PDF+ uncertainty. These varying- PDF sets can be constructed from a simple average of existing sets from the individual groups. If we denote by the central value of combined PDF4LHC set for a specific quark flavor (or the gluon), then the sets with different values of can be constructed as follows:
and
with and the corresponding central values of the MMHT14, CT14 and NNPDF3.0 sets for those specific values of We have verified that other possible ways of constructing these sets (such as different interpolation options) do not change the result in any appreciable way.
5.2. The MC reduction method: compressed Monte Carlo PDFs (CMC-PDFs)
CMC-PDFs [113] are determined by using a compression algorithm, that, starting from an MC prior with replicas, determines the set of replicas that most faithfully reproduce the original probability distribution in terms of central values, variances, higher moments and corrections. Therefore with the CMC-PDFs one ends up with an MC representation of the original MC900 combination but based on a much reduced number of replicas (about an order of magnitude reduction), with minimal information loss.
The compression algorithm is based on the minimization of a figure of merit
where k runs over the number of statistical estimators used to quantify the distance between the original and compressed distributions, Nk is a normalization factor, is the value of the estimator k (for example, the mean or the variance) computed at the generic point and is the corresponding value of the same estimator in the compressed set. The various contributions to equation (3) include the mean, variance, skewness, kurtosis, the Kolmogorov distance and the correlations between PDFs. The minimization is performed using Genetic Algorithms. The main advantage of the CMC-PDF method is that not only the central value and variance of the original distribution are reproduced, but also its higher moments: this is of course crucial when the underlying probability distribution is non-Gaussian, as illustrated in figures 10 and 11.
Download figure:
Standard image High-resolution imageThe quality of the compression has been extensively validated in [113] at the level of PDFs and of LHC cross-sections, including their correlations; the general conclusion is that replicas are enough to reproduce all relevant statistical estimators of the original combined PDF set and can be reliably used in LHC phenomenology: a comparative quantitative assessment will be given below.
The choice of estimators included in the figure of merit equation (3) is a compromise between the goal of reproducing reasonably well both the low moments, and specifically the mean and standard deviation of the distributions (which fully determine the Gaussian approximation), and the higher moments of the distribution. Specifically, if only the Kolmogorov distance were included, all moments would be reproduced equally well, while if only the mean and standard deviation were included, only these would be optimized. Therefore, in general there is a trade-off in accuracy: the CMC-PDFs will perform slightly worse in reproducing the central values and variances of the prior as compared to a Hessian reduction, but the advantage of keeping the non-Gaussian features which are missed by the latter case.
The CMC100 PDFs have been used to construct the default MC representation of the PDF4LHC combined sets that are made available on LHAPDF6:
In addition, the compressor code use to generate the CMC-PDF sets is also publicly available from
https://github.com/scarrazza/compressor
together with the corresponding user documentation. For completeness, section 6.2 reports the formulae that should be used to compute the PDF and PDF+ uncertainties whenever the PDF4LHC15_mc sets are used.
5.3. Hessian reduction methods
A Hessian representation of PDF uncertainties has certain advantages and disadvantages in comparison to an MC representation. The main disadvantage is that a Hessian representation assumes that the underlying probability distribution is multi-Gaussian (though in general possibly asymmetric, i.e. with different upper and lower uncertainties). The main advantage is that Hessian uncertainties can be treated as nuisance parameters, and thus on the same footing as other nuisance parameters: this can be useful in analysis, for example for the determination of PDF-induced correlations in large datasets, and also when using profiling in order to understand the dominant PDF contributions to a given process.
Two techniques have been suggested to turn an MC PDF set into a Hessian set: META-PDF and MCH-PDFs. By carefully tuning the number of Hessian sets which are used, these techniques also provide a suitable reduction method. Clearly, when choosing the number of eigenvectors to be included in a Hessian representation of PDF errors there is a trade-off between speed and accuracy. The optimal number of thus depends then on the specific application, i.e., whether speed or accuracy is the most important consideration. We will now present each of the two methods, and then discuss the choice of the number of PDF error members.
5.3.1. META-PDFs
The meta-analysis of [111] was the first method proposed for reducing the number of MC replicas in the combined PDF ensemble. The META analysis starts from the MC representation of the input PDF set. Then, each MC replica is re-fitted using a flexible 'meta-parametrization', so that all input replicas are cast into a common parametrized form. The probability distribution is examined as a function of parameters ai of the meta-parametrizations. The least and best constrained combinations of ai are found by diagonalization of the covariance matrix on the PDF parameter space.
From this information, one can construct Hessian error PDFs that reproduce intervals of given confidence along each poorly constrained direction, centered on the average PDF set of the input ensemble. In contrast, error PDFs corresponding to displacements along well-constrained directions contribute little to the total PDF uncertainty and can be discarded. In the end, one obtains a central PDF, corresponding to an average of the input replicas; as well as an ensemble consisting of a relatively small number of Hessian eigenvectors, reproducing the principal components of the original covariance matrix.
Introduction of meta-parametrizations thus leads to a fairly intuitive method for combination, which reduces to averaging of discrete parameters and diagonalization of a covariance matrix in a quasi-Gaussian approximation. While the choice of the functional form for the meta-parametrizations is not unique, it does not bias the reduced META-PDFs in practice, which has been verified by trying various parametrization forms and fitting procedures.
In the current study, we construct a META-PDF combination with one central and symmetric eigenvectors according to the following procedure. We start from the same prior MC900, see section 5.1 as done in the other reduction methods. Each PDF replica is then re-fitted, at a given input scale Q0, using the form
where is the average PDF set for flavor α. The basis functions are
for and include Bernstein polynomials of degree
The parameters are zero for the central PDF set, but vary for each of the MC replicas. They are chosen to minimize a metric function E on a grid of momentum fraction values in the interval at the initial scale Q0:
where and is the symmetric PDF uncertainty of
The value of Q0 where the MC replicas are parametrized is taken to be well above the bottom-quark mass mb, where all input heavy-quark schemes lead to close predictions. Thus we fit parameters for 9 independent PDF flavors at Q0, and setting for simplicity. Small differences between sea quarks and antiquarks induced purely by NNLO evolution effects are negligible here, given that Q0 is still relatively low.
The covariance matrix in the space of PDF parameters is computed according to
and diagonalized by an orthogonal transformation O:
The eigenvalues of the covariance matrix are positive-definite. The eigenvectors are associated with the parameters We can then reasonably interpret to be the width of the effective Gaussian distribution describing and neglecting any asymmetry of this distribution.
To construct the META ensemble we select the largest, or principal, eigenvalues and discard the eigenvectors associated with well-constrained For the discarded directions, the respective are set to zero, their average value; the number of eigenvectors for the final META-PDFs thus reduces below the original number of 144. Using the META representation, a standard deviation on a quantity X can be computed using the usual Hessian symmetric master formula,
where and are the predictions for the central and -th eigenvector of the META ensemble. This is a symmetrized master formula, which provides an estimate of the 68% CL uncertainty at all Q values, providing the DGLAP evolution of all input PDFs is numerically consistent17 . In this approach, common DGLAP evolution settings are used for all the META eigenvectors using the HOPPET program [105].
The Hessian META uncertainty provides a lower estimate for the MC900 uncertainty. For example, the META-PDF set with symmetric eigenvectors (denoted by META100 in the following) exhibits PDF uncertainties closer to the MC900 uncertainty than the META30 uncertainty (the META-PDF set with symmetric eigenvectors), but requires the evaluation of a larger number of eigenvectors. We find that the META30 combination efficiently captures information about the lowest two (Gaussian) moments of the original MC900 distribution, which are the most robust and predictable. The number corresponds to about the minimal number of parameters that are needed to describe PDF degrees of freedom in at least three independent x regions (small, intermediate, and large x) for the nine physical flavors.
In a sense, META30 provides a 'minimal' estimate of the PDF uncertainty, so that 'the true uncertainty must be at least as much with good confidence'. META100, on the other hand, provides a more complete estimate for the given prior, by including subleading contributions that are also more prone to variations.
The META30 PDFs are parametrized at a starting scale of Q0 = 8 GeV, above which one can neglect differences in the treatment of heavy-quark flavors in the CT14, MMHT2014, and NNPDF3.0 PDF sets. This is sufficient for computing hard-scattering cross sections for typical LHC observables using the Nf = 5 massless approximation. However, in those calculations that match the finite-order cross sections with initial-state parton showers, knowing the PDFs at scales below 8 GeV may also be necessary in order to calculate the Sudakov form factors. For these specific applications, the META30 PDFs have been extended down to Qmin = 1.4 GeV by backward DGLAP evolution, and the stability of this low-Q extension was validated using several tests. The resulting uncertainties at low Q are in good agreement with those from the MCH-100 PDFs. In the publicly provided LHAPDF grids for the PDF4LHC1530 set, the whole range of Q ≥ 1.4 GeV is covered; the transition across the fitting scale of 8 GeV is invisible to the user. Only above 8 GeV the treatment of heavy flavors is fully consistent; below this scale the combined PDFs are suitable for specific calculations, such as those involving parton showering generators.
5.3.2. MCH-PDFs
The mc2Hessian algorithm for construction of a Hessian representation of an MC PDF set uses the replicas themselves as a linear expansion basis. This idea was realized in two different ways in [112]. The first method first identifies the region where the Gaussian approximation is valid, and uses Genetic Algorithms to identify the optimal expansion basis. The second method uses SVD to represent the covariance matrix on a basis of replicas.
This second method is the preferred one if the goal is achieving an optimal reproduction of the covariance matrix, and it is the method which is reviewed here. In this method, the multi-Gaussian distribution of replicas is viewed as a distribution of deviations of PDFs from their central value. If independent PDFs at scale Q are sampled on a fine enough grid of Nx points in xi, the deviations from the central PDF can be collected in a -dimensional vector and the multi-Gaussian is fully determined by the corresponding covariance matrix. This, in turn, is represented as a linear combination of the starting replicas ,
This is done in the following way. The set of vectors X of deviations from the central value, for each replica
where α labels PDFs, i points in the x grid, runs over all grid points, k runs over all the MC replicas, and can be viewed as a rectangular matrix. The covariance matrix can then be written as
and therefore
A representation of the covariance matrix in terms of replicas is found by SVD of the matrix X. Namely, we can write X as:
Assuming that U is an orthogonal matrix of dimensions which contains the orthogonal eigenvectors of the covariance matrix with nonzero eigenvalues; S is a diagonal matrix of real positive elements, constructed out of the singular values of X, i.e., the square roots of the nonzero eigenvalues of the covariance matrix equation (12), multiplied by the normalization constant and V is an orthogonal matrix of coefficients.
Because
the matrix Z = US has the property that
But also
and thus Z provides the sought-for representation of the covariance matrix as a linear combination of MC replicas.
Note that by 'linear combination of replicas' here we mean that each eigenvector is a combination of the original PDF replicas i.e., with coefficients which depend on the replica index k, but do not depend on either the PDF index α or the value of x; nor, because of linearity of QCD evolution, on the scale Q. The method therefore allows for combination of PDFs that evolve in slightly different ways, with results that do not depend on the scale at which the SVD is performed.
A reduction of the number of eigenvectors can now be performed using PCA. Since we observe that in practice many of the eigenvectors lead to a very small contribution to the covariance matrix, we can select a smaller set of eigenvectors which still provides a good approximation to the covariance matrix by selecting the eigenvectors with largest eigenvalues.
Denoting with u, s, and v the , and reduced matrices computed using these eigenvalues, for a given value of using v instead of V in equation (17) minimizes the difference between the original and reduced covariance matrix
This method is guaranteed by construction to reproduce the covariance matrix of the initial prior set, that is, both PDF uncertainties and PDF correlations with an arbitrary level of precision, upon including more eigenvectors. It is also numerically stable (in the sense that the SVD yields well conditioned matrices), and the same methodology can be applied no matter which input PDF sets are used in the MC combination.
In practice, we add eigenvectors until the size of the differences between the new Hessian and the original MC representations becomes comparable to the accuracy of the Gaussian approximation in that specific region. This can be characterized by the difference between the and 68% CL intervals of the distribution of results when computing LHC cross-sections prior MC set MC900. The interval will overestimate the effect of the outlier replicas in comparison with the 68% CL intervals. Note that insisting that the method reproduces the band exactly would actually lead to bigger differences between the prior and the Hessian distribution, as measured for example by the Kolmogorov distance between both. Therefore we add eigenvectors until the Hessian prediction for the ends up between the 68% CL and the intervals of the prior MC900.
As will be discussed in the comparison section below, we find that in the mc2hessian algorithm provides a satisfactory description of both the data regions and the extrapolation regions can be realized with symmetric eigenvectors, for all relevant of observables. Decreasing this number does not seem advisable, since then one needs to provide ad hoc assumptions to increase the correlation lengths of the points in the small and large-x regions.
5.3.3. Choice of Hessian sets
Extensive benchmarking of the two available Hessian methods has been performed, both at the level of parton distributions, luminosities, and LHC cross-sections. The complete set of comparison plots can be found at the PDF4LHC website: http://hep.ucl.ac.uk/pdf4lhc/mc2h-gallery/website.
It turns out that for a low number of eigenvectors, of the order of META-PDFs has a somewhat superior performance, leading to a generally satisfactory representation of the PDF covariance matrix except in outlying regions, where the META-PDF uncertainties are smaller than those of the MC900 prior because of its minimal nature (see section 5.3.1). On the other hand, for a larger number of eigenvectors, the MCH-PDFs yield an extremely accurate representation of the covariance matrix with a somewhat improved performance as compared to the META-PDF method. For instance, as will be shown below, the entries of the PDF correlation matrix can be reproduced at the per-mile level.
For these reasons, in the following the Hessian PDF4LHC15 combined sets with eigenvectors
are constructed from the META30 sets, while those with eigenvectors
are constructed from the MCH100 sets.
The MP4LHC code used to generate the META sets is publicly available from the HepForge website: https://metapdf.hepforge.org/.
The mc2Hessiancode use to generate the MC2H sets is also publicly available from: https://github.com/scarrazza/mc2Hessian together with the corresponding user documentation. For completeness, section 6.2 reports the formulae that should be used to compute the PDF and PDF+ uncertainties whenever any of the two Hessian sets, PDF4LHC15_100 and PDF4LHC15_30, are used.
5.4. Comparisons and benchmarking
Now we compare the three combined PDF sets resulting from applying different reduction strategies to the MC900 prior: the MC set PDF4LHC15_mc with replicas, and the two Hessian sets PDF4LHC15_100 and PDF4LHC15_30, with and 30 symmetric eigenvectors respectively. We will only show here some representative comparison plots between the three NNLO sets, with the complete set of plots of PDFs, luminosities and LHC cross-sections (as well as the corresponding comparisons at NLO) are available from the PDF4LHC website: http://hep.ucl.ac.uk/pdf4lhc/mc2h-gallery/website.
We start the comparisons of the prior with the various reduced sets at the level of PDFs, which we perform at a representative LHC scale of GeV2. In figure 12 we compare the prior and the MC compressed set PDF4LHC15_nnlo_mc. We show results for the gluon, total quark singlet, anti-up quark and the isoscalar triplet, normalized to the central value of PDF4LHC15_nnlo_prior. As can be seen, there is good agreement in all the range in x both for central values and for PDF uncertainties.
Download figure:
Standard image High-resolution imageThe corresponding comparison between the prior MC900 and the two reduced Hessian sets, PDF4LHC15_nnlo_100 and PDF4LHC15_nnlo_30, is shown in figure 13. In the case of the set with eigenvectors, the agreement is good for all PDF combinations in the complete range of x. For the case of the set with eigenvectors, the agreement is also good in the PDF x range corresponding to precision physics measurements, but renders a slightly smaller uncertainty in the extrapolation regions at small- and large-x. Also for PDF combinations that are not known very well, like the isoscalar triplet, again the uncertainty of sets can be below that of the prior in the poorly constrained regions.
Download figure:
Standard image High-resolution imageNow we turn to a comparison of parton luminosities. In figure 14 we compare the parton luminosities at the LHC 13 TeV computed with the prior set PDF4LHC15_nnlo_prior, with and its CMC representation, PDF4LHC15_nnlo_mc, for We show the gg, qg, qq and luminosities as a function of the invariant mass of the final state MX, normalized to the central value of PDF4LHC15_nnlo_prior. We see reasonable agreement in all cases.
Download figure:
Standard image High-resolution imageThe corresponding comparison of PDF luminosities in the case of the reduced Hessian sets is shown in figure 15, where we compare the prior set PDF4LHC15_nnlo_prior and the two Hessian sets, PDF4LHC15_nnlo_100 and PDF4LHC15_nnlo_30. Both the set and the provide a good representation of the 900 replica prior in the region of MX relevant for precision physics while the the performs slightly worse in the low mass region, sensitive to the small-x PDFs.
Download figure:
Standard image High-resolution imageA more detailed quantification of the differences in the three reduction methods can be obtained by computing the percentage difference of variances between the prior and its various reduced representations. Recall that in the Hessian approach, central values are reproduced automatically, as is the deviation from the central set which is parametrized by the error sets; in the MC approach in principle, there can be a small shift in the mean when reducing the sample size, though this is of the order of the standard deviation of the mean (i.e. 10 time smaller than the standard deviation, for a 100-replica sample), further reduced by the compression method. We have explicitly checked that indeed in all cases the accuracy with which central values are reproduced is that of the LHAPDF6 interpolation with which final grids are delivered.
Results for the percentage differences in the variance (that is, in the PDF uncertainties) between the prior and each of the three reduced sets are shown in figure 16. We see that Hessian set with 100 eigenvectors reproduces in all cases the variances of the prior with precision at worst of 1%, typically much better than that. The MC set and the Hessian with 30 eigenvectors lead to similar performances in terms of reproducing the variances of the prior, with differences that can be up to 10%.
Download figure:
Standard image High-resolution imageWe then turn to a comparison of PDF correlations. In figure 17 we show the difference between the PDF correlation coefficients computed with the prior set PDF4LHC15_nnlo_prior and its CMC representation, PDF4LHC15_nnlo_mc. The plot shows the 7 × 7 matrix of all possible correlations of light flavor PDFs (three quarks and antiquarks and the gluon), sampled on logarithmically spaced x values between and and with Q = 8 GeV. In the left plot of figure 17 the range of difference shown is between −1 and 1, while in the right plot we zoom in to a range of difference between −0.2 and 0.2. We find good agreement in general, with differences never exceeding 0.2 in modulus. Note that the correlation coefficients can easily vary by 0.1–0.2 or more among the input PDF sets.
Download figure:
Standard image High-resolution imageThe corresponding comparison of PDF correlations for the two Hessian sets, PDF4LHC15_nnlo_100 (upper plots) and PDF4LHC15_nnlo_30 (lower plots), is shown in figure 18. For the case of the set the correlations are essentially identical to those of the prior, with differences below 0.01. For the set instead, typical differences in the correlation coefficients are somewhat larger, comparable to those found when using the MC set PDF4LHC15_nnlo_mc in figure 17.
Download figure:
Standard image High-resolution image5.5. Implications for LHC processes
We now consider the comparison at the level of LHC cross-sections and distributions. All comparisons reported in this document have been performed using NLO matrix elements, for which fast computational tools are available, even when using NNLO PDFs: the comparison is performed for PDF validation purposes only. The NLO calculations that have been used for the validation presented in this report have been produced used APPLgrid [90] interfaced to NLOjet++ [89] and MCFM [15], and with aMCfast [117] interfaced to Madgraph5_aMC@NLO [118]. In all cases we use the default theory settings, including the scale choices, of the respective codes. Theory calculations have been performed at 7 TeV for these processes for which data is already available and that have been used for PDF fits, see for example [11]; in addition, a number of dedicated grids for 13 TeV processes have also been generated. In the former case the binning follows that to the corresponding experimental measurements, which we indicate in the list below.
In this report we will only show a representative subset of processes. The complete list of processes and kinematic distributions (available from the PDF4LHC15 webpage) is described now. At 7 TeV, the processes where experimental LHC measurements are available, and for which APPLgrid grids matching the experimental binning have been produced, are the following:
- DY rapidity distributions in the LHCb forward region [36, 119].
- Invariant mass distribution from high-mass DY [120].
- Rapidity distributions of W and Z production [34, 50].
- pT distribution of inclusive W production [121].
- Double differential DY distributions in dilepton mass and rapidity [122].
- Lepton rapidity distributions in W + charm production [116].
- Inclusive jet production in the central and forward regions [33, 123].
In addition to these grids, at 13 TeV we have generated specifically for this benchmark exercises a number of new fast NLO grids for differential distributions in Higgs, and vector boson production using Madgraph5_aMC@NLO interfaced to aMCfast, namely:
- Rapidity and pT distributions in inclusive production, as well as total cross-sections for hZ, hW and production.
- Rapidity, pT and distributions in top-quark pair production.
- Missing ET, lepton pT and rapidity, and transverse mass distributions in inclusive W and Z production.
First of all, let us go back to figures 10 and 11, which illustrated two cases of LHC cross-sections where departures from the Gaussian regime were particularly striking. In these figures, the probability distributions computed from the MC900 prior, is compared to those computed using the CMC100 (PDF4LHC15_mc) and MCH100 (PDF4LHC15_100) reduced sets. It is clear that while the MC PDF4LHC15_mc set is able to capture these non-Gaussian features, this is not the case for the Hessian set PDF4LHC15_100.
Then, in figure 19 we show the comparison between the three PDF4LHC15 combined sets for a representative selection of LHC cross-sections. In particular we show from top to bottom and from left to right the forward DY rapidity distributions at LHCb, the CMS DY double-differential distributions, the ATLAS 2010 inclusive jets in the central and forward regions (all these at 7 TeV) and then the pT and rapidity distributions of for Higgs production in gluon fusion, the distribution in production and the Z pT distribution (all these at 13 TeV). We compare the prior with the three reduced sets, the MC and the two Hessian sets. As can be seen, the agreement is in general good in all cases.
Download figure:
Standard image High-resolution imageAn important application of the PDF4LHC15 combined sets is the computation of the correlation coefficients between LHC processes, such as for instance signal and background processes in Higgs production. Within the HXSWG, the correlation coefficients between some of these processes were computed using the original PDF4LHC recommendation: see table 10 of [25]. We have recomputed these correlation coefficients using MC900, and compared the results with the three reduction methods. The processes included, at the LHC 13 TeV, are vector boson production, both W and Z, production, and then Higgs production in gluon fusion ggh, in associated production hZ and hW, and in association with a top quark pair,
In figure 20 we show the difference between the correlation coefficients computed using the prior set PDF4LHC15_nnlo_prior and the two Hessian sets, PDF4LHC15_nnlo_100 (upper left plot) and PDF4LHC15_nnlo_30 (upper right plot) as well as with the MC set, PDF4LHC15_nnlo_mc (lower plot). As we can see, with the set the correlation coefficients are always reproduced within a few percent, while for the and the MC sets somewhat larger differences are found, a few instances up to 0.2.
Download figure:
Standard image High-resolution imageIn addition, we have tabulated the correlation coefficients for various LHC total cross-sections computed using the prior and the various reduced sets. First of all, in table 1 we show the value of the correlation coefficient between the Z production cross-sections and the W, , ggh, , hW, and hZ production cross-sections. We compare the PDF4LHC15 prior with the MC and the two Hessian reduced sets, both at NLO and at NNLO.
Table 1. Correlation coefficient between the Z production cross-sections and the W, , ggh, , hW and hZ production cross-sections. The PDF4LHC15 prior is compared to the Monte Carlo and the two Hessian reduced sets, both at NLO and at NNLO.
Correlation coefficient | ||||||
---|---|---|---|---|---|---|
PDF set | ||||||
PDF4LHC15_nlo_prior | 0.90 | −0.60 | 0.22 | −0.64 | 0.55 | 0.74 |
PDF4LHC15_nlo_mc | 0.92 | −0.49 | 0.41 | −0.58 | 0.61 | 0.77 |
PDF4LHC15_nlo_100 | 0.92 | −0.60 | 0.23 | −0.64 | 0.57 | 0.75 |
PDF4LHC15_nlo_30 | 0.90 | −0.68 | 0.16 | −0.71 | 0.55 | 0.76 |
PDF4LHC15_nnlo_prior | 0.89 | −0.49 | 0.08 | −0.46 | 0.56 | 0.74 |
PDF4LHC15_nnlo_mc | 0.90 | −0.44 | 0.18 | −0.42 | 0.62 | 0.80 |
PDF4LHC15_nnlo_100 | 0.91 | −0.48 | 0.09 | −0.46 | 0.59 | 0.74 |
PDF4LHC15_nnlo_30 | 0.88 | −0.63 | 0.04 | −0.61 | 0.56 | 0.72 |
The information from this table is the consistent with that shown in figure 20, namely the Hessian set with 100 eigenvectors reproduces always the correlation coefficient of the prior, with an accuracy better than 1%. Using the smaller Hessian set and the MC set the correlation coefficients are reproduced not as perfectly, but still sufficiently well for most phenomenological applications. It should be kept in mind that the correlations themselves from the prior set have sizable uncertainties, in the sense that the individual correlations from the three different PDF families may differ significantly from each other. Thus, the correlations given in the tables provide an average behavior over the three PDF sets, and probably only the first digit in the correlation coefficient should be considered significant. The same conclusion can be derived from a variety of other pairs of LHC cross-sections, collected in tables 2–4.
Table 2. Same as table 1 for the correlation coefficient of additional pairs of LHC inclusive cross-sections.
Correlation coefficient | ||||||
---|---|---|---|---|---|---|
PDF set | ||||||
PDF4LHC15_nlo_prior | −0.46 | 0.32 | −0.51 | 0.77 | 0.78 | 0.27 |
PDF4LHC15_nlo_mc | −0.35 | 0.49 | −0.46 | 0.81 | 0.80 | 0.27 |
PDF4LHC15_nlo_100 | −0.47 | 0.32 | −0.52 | 0.77 | 0.79 | 0.27 |
PDF4LHC15_nlo_30 | −0.52 | 0.28 | −0.56 | 0.79 | 0.81 | 0.32 |
PDF4LHC15_nnlo_prior | −0.40 | 0.20 | −0.40 | 0.76 | 0.77 | 0.30 |
PDF4LHC15_nnlo_mc | −0.44 | 0.26 | −0.42 | 0.81 | 0.82 | 0.32 |
PDF4LHC15_nnlo_100 | −0.40 | 0.20 | −0.40 | 0.76 | 0.77 | 0.30 |
PDF4LHC15_nnlo_30 | −0.47 | 0.19 | −0.47 | 0.77 | 0.76 | 0.31 |
Table 3. Same as table 1 for the correlation coefficient of additional pairs of LHC inclusive cross-sections.
Correlation coefficient | ||||||
---|---|---|---|---|---|---|
PDF set | ||||||
PDF4LHC15_nlo_prior | 0.93 | −0.22 | −0.50 | −0.02 | 0.15 | 0.08 |
PDF4LHC15_nlo_mc | 0.92 | −0.14 | −0.41 | −0.04 | 0.33 | 0.27 |
PDF4LHC15_nlo_100 | 0.93 | −0.22 | −0.48 | −0.03 | 0.15 | 0.08 |
PDF4LHC15_nlo_30 | 0.93 | −0.25 | −0.54 | 0.02 | 0.11 | −0.01 |
PDF4LHC15_nnlo_prior | 0.87 | −0.23 | −0.34 | −0.13 | −0.01 | −0.17 |
PDF4LHC15_nnlo_mc | 0.87 | −0.27 | −0.35 | −0.10 | 0.07 | −0.01 |
PDF4LHC15_nnlo_100 | 0.87 | −0.24 | −0.34 | −0.13 | −0.02 | −0.17 |
PDF4LHC15_nnlo_30 | 0.87 | −0.27 | −0.43 | −0.13 | −0.04 | −0.23 |
Table 4. Same as table 1 for the correlation coefficient of additional pairs of LHC inclusive cross-sections.
Correlation coefficient | |||
---|---|---|---|
PDF set | |||
PDF4LHC15_nlo_prior | −0.18 | −0.43 | 0.88 |
PDF4LHC15_nlo_mc | −0.15 | −0.41 | 0.87 |
PDF4LHC15_nlo_100 | −0.18 | −0.42 | 0.89 |
PDF4LHC15_nlo_30 | −0.19 | −0.46 | 0.88 |
PDF4LHC15_nnlo_prior | −0.13 | −0.17 | 0.90 |
PDF4LHC15_nnlo_mc | −0.17 | −0.21 | 0.90 |
PDF4LHC15_nnlo_100 | −0.13 | −0.17 | 0.91 |
PDF4LHC15_nnlo_30 | −0.17 | −0.25 | 0.91 |
6. The PDF4LHC 2015 recommendations
The 2015 PDF4LHC prescription has been motivated, and its general principles spelled out, in section 4. As discussed there, while in some cases individual PDF sets should always be used, for other LHC applications a combination of PDF sets is required. For these cases, in section 5 we have constructed the statistical combination of the CT14, MMHT14 and NNPDF3.0 PDF sets using the MC method. This strategy is different from the envelope method employed in the 2010 recommendations, and allows a robust statistical interpretation of the ensuing PDF uncertainties
The PDF4LHC15 combined sets are then delivered using three different options, corresponding in all cases to the same underlying prior combination. The choice of which delivery method should be used depends on purely practical considerations, such as whether a Hessian or MC representation is preferred, or on the trade-off between computational speed and accuracy. Explicit recommendations for the usage of each of these three delivery options are provided below.
In this section we first present the final PDF4LHC15 sets which will be made available in LHAPDF6 and provide general guidelines for their usage. Then we review the formulae for the calculation of PDF and PDF + uncertainties in each case. Finally, we present the PDF4LHC15 combined sets to be used in calculations in the nf = 4 scheme, and summarize the citation policy that should be used whenever these combined PDF sets are used.
We emphasize that the present document contains recommendations, not a series of unique instructions. We believe that with the delivery options provided here any PDF user will find the flexibility to choose the strategy that is better suited for each particular analysis.
6.1. Delivery and guidelines
The PDF4LHC15 combined PDFs are based on an underlying MC combination of CT14, MMHT14 and NNPDF3.0, denoted by MC900, which is made publicly available in three different reduced delivery forms:
- PDF4LHC15_mc: an MC PDF set with replicas.
- PDF4LHC15_30: a symmetric Hessian PDF set with eigenvectors.
- PDF4LHC15_100: a symmetric Hessian PDF set with eigenvectors.
In the three cases, combined sets are available at NLO and at NNLO, for the central value of In addition, we provide additional sets which contain the central values for and and that can be used for the computation of the combined PDF + uncertainties, as explained in section 6.2. Finally, for ease of usage, the combined sets for are also presented bundled with the -varying sets in dedicated grid files. The specifications of each of the combined NNLO PDF4LHC15 sets that are available from LHAPDF6 are summarized in table 5; note that the corresponding NLO sets are also available.
Table 5. Summary of the combined NNLO PDF4LHC15 sets with that are available from LHAPDF6. The corresponding NLO sets are also available. Members 0 and 1 of PDF4LHC15_nnlo_asvar coincide with members 101 and 102 (31 and 32) of PDF4LHC15_nnlo_mc_pdfas and PDF4LHC15_nnlo_100_pdfas (PDF4LHC15_nnlo_30_pdfas). Recall that in LHAPDF6 there is always a zeroth member, so that the total number of PDF members in a given set is always See text for more details.
LHAPDF6 grid | Pert order | ErrorType | ||
---|---|---|---|---|
PDF4LHC15_nnlo_mc | NNLO | replicas | 100 | 0.118 |
PDF4LHC15_nnlo_100 | NNLO | symmhessian | 100 | 0.118 |
PDF4LHC15_nnlo_30 | NNLO | symmhessian | 30 | 0.118 |
PDF4LHC15_nnlo_mc_pdfas | NNLO | replicas+as | 102 | mem 0:100 → 0.118 |
mem 101 → 0.1165 | ||||
mem 102 → 0.1195 | ||||
PDF4LHC15_nnlo_100_pdfas | NNLO | symmhessian+as | 102 | mem 0:100 → 0.118 |
mem 101 → 0.1165 | ||||
mem 102 → 0.1195 | ||||
PDF4LHC15_nnlo_30_pdfas | NNLO | symmhessian+as | 32 | mem 0:30 → 0.118 |
mem 31 → 0.1165 | ||||
mem 32 → 0.1195 | ||||
PDF4LHC15_nnlo_asvar | NNLO | — | 1 | mem 0 → 0.1165 |
mem 1 → 0.1195 |
Usage of the PDF4LHC15 sets. As illustrated in section 5, the three delivery options provide a reasonably accurate representation of the original prior combination. However, each of these methods has its own advantages and disadvantages, which make them more suited in different specific contexts. We now attempt to provide some general guidance about which of the three PDF4LHC15 combined sets should be used in specific phenomenological applications.
- (1)Comparisons between data and theory for standard model measurementsRecommendations: Use individual PDF sets, and, in particular, as many of the modern PDF sets [5–11] as possible.Rationale: Measurements such as jet production, vector-boson single and pair production, or top-quark pair production, have the power to constrain PDFs, and this is best utilized and illustrated by comparing with many individual sets.As a rule of thumb, any measurement that potentially can be included in PDF fits falls in this category.The same recommendation applies to the extraction of precision SM parameters, such as the strong coupling [75, 124], the W mass MW [125], and the top quark mass mt [126] which are directly correlated to the PDFs used in the extraction.
- (2)Searches for beyond the standard model phenomenaRecommendations: Use the PDF4LHC15_mc sets.Rationale: BSM searches, in particular for new massive particles in the TeV scale, often require the knowledge of PDFs in regions where available experimental constraints are limited, notably close to the hadronic threshold where [127]. In these extreme kinematical regions the PDF uncertainties are large, the MC combination of PDF sets is likely to be non-Gaussian. See figures 10 and 11.This case also applies to the calculations of PDF uncertainties in related theoretical analysis, such as when determining exclusion limits for specific BSM scenarios.If it is necessary to use a Hessian representation of the PDF uncertainty, for example in order to express PDF errors as Gaussian systematic uncertainties, one can cross-check the PDF4LHC15_mc results with the two Hessian sets, PDF4LHC15_30 and PDF4LHC15_100, depending on the required accuracy.
- (3)Calculation of PDF uncertainties in situations when computational speed is needed, or a more limited number of error PDFs may be desirableRecommendations: Use the PDF4LHC15_30 sets.Rationale: In many situations, PDF uncertainties may affect the extraction of physics parameters. From the point of view of the statistical analysis, it might be useful in some cases to limit the number of error PDFs that need to be included in such analyzes. In these cases, use of the PDF4LHC15_30 sets may be most suitable.In addition, the calculation of acceptances, efficiencies or extrapolation factors are affected by the corresponding PDF uncertainty. These quantities are only a moderate correction to the measured cross-section, and thus a mild loss of accuracy in the determination of PDF uncertainties in these corrections is acceptable, while computational speed can be an issue. In these cases, use of the PDF4LHC15_30 sets is most suitable.However, in the cases when PDF uncertainties turn out to be substantial, we recommend to cross-check the PDF estimate by comparing with the results of the PDF4LHC15_100 sets.
- (4)Calculation of PDF uncertainties in precision observablesRecommendation: Use the PDF4LHC15_100 sets.Rationale: For several LHC phenomenological applications, the highest accuracy is sought for, with, in some cases, the need to control PDF uncertainties to the percent level, as currently allowed by the development of high-order computational techniques in the QCD and electroweak sectors of the Standard Model.Whenever the highest accuracy is desired, the PDF4LHC15_100 set is most suitable.However, calculations that have little impact on the PDF dependence of the precision measurement, such as certain acceptances, may also be computed according to (3), using the PDF4LHC15_30 set.
Concerning the specific applications of the four cases listed above, there are some important caveats to take into account.
- For the same process, more than one of the user cases above might be applicable. For instance, consider the total top quark production cross-section one should use either (3) or (4) to estimate the total PDF uncertainty (for instance to determine the overall compatibility of the SM theory and data for this measurement), and at the same time use (1) to gauge the sensitivity of this observable to constrain PDFs.Therefore, cases (1)–(4) above are not exclusive: one or the other should be more adequate depending on the theoretical interpretation of a given experimental measurement.
- Since the three delivery methods are based on the same underlying PDF combination, it is perfectly consistent to use different delivery options in different parts of the same calculation.For instance, for the computation of the production cross-sections of high-mass BSM resonances, one could use the PDF4LHC15_mc sets to estimate the PDF uncertainty in the expected signal yield, and at the same time the PDF4LHC15_30 sets to estimate the PDF uncertainties that affect the corresponding acceptance calculation.
- In the case of a really significant discrepancy between the experimental data and theoretical calculations, it may be crucial to fully exclude the possibility that this discrepancy arises due to the PDFs.In this case we recommend to compare data with a broader variety of PDFs, including PDF sets without LHC data, or PDF sets based only on DIS data.
6.2. Formulae for the calculation of PDF and PDF+ uncertainties
For completeness, we also collect in this report the explicit formulae for the calculation of PDF and combined PDF+ uncertainties in LHC cross-sections when using the PDF4LHC15 combined sets. Let us assume that we wish to estimate the PDF+ uncertainty of given cross-section σ, which could be a total inclusive cross-section or any bin of a differential distribution.
First of all, to compute the PDF uncertainty, one has to evaluate this cross-section times, where is the number of error sets (either symmetric eigenvectors or MC replicas) of the specific combined set
so in particular in PDF4LHC15_30 and in PDF4LHC15_100 and PDF4LHC15_mc.
PDF uncertainties for Hessian sets. In the case of the Hessian sets, PDF4LHC15_30 and PDF4LHC15_100, the master formula to evaluate the PDF uncertainty is given by
This uncertainty is to be understood as a 68% CL. From this expression it is also easy to determine the contribution of each eigenvector k to the total Hessian PDF uncertainty.
PDF uncertainties for MC sets. For the case of the MC sets, PDF4LHC15_mc, PDF uncertainties can be computed in two ways. First of all, one can use the standard deviation of the distribution
where the mean value of the cross-section, is computed as usual
For MC sets one should use always equation (22) for the mean value of the cross-section, though in many cases it is true that
Alternatively, PDF uncertainties in an MC set can be computed from the 68% CL. This is achieved by reordering the values for the cross-sections in ascending order, so that we have
and then the PDF uncertainty computed as the 68% CL interval of will be given by
This definition is suitable wherever departures from the Gaussian regime are sizable, since it gives the correct statistical weight to outliers. In general, it can be useful to compare results with the two expressions, equations (21) and (24), and whenever differences are found, use equation (24). In addition, in the non-Gaussian case the mean of the distribution equation (22) is not necessarily the best choice for the central value of the cross-section, and we recommend to use instead the midpoint of the 68% CL interval, that is
Combined uncertainties. Let us now turn to discuss the computation of the combined PDF+ uncertainties. The PDF4LHC15 combined are based on the following value of and of its associated uncertainty,
at the 68% CL, and both at NLO and at NNLO. This choice is consistent with the current PDG average [107], and reflects recent developments towards the updated 2015 PDG average18 . It is then recommended that PDF+ uncertainties are determined by first computing the PDF uncertainty for the central using equations (20), (21) or (24), then computing predictions for the upper and lower values of consistently using the corresponding PDF sets, and finally adding results in quadrature.
Specifically, for the same cross-section σ as before, the uncertainty can be computed as:
corresponding to an uncertainty at the 68% confidence level. Note that equation (27) is to be computed with the central values of the corresponding PDF4LHC15 sets only. Needless to say, the same value of should always be used in the partonic cross-sections and in the PDFs. The combined PDF+ uncertainty is then computed as follows
The result for any other value of as compared to the baseline equation (26), can be obtained from a trivial rescaling of equation (27) assuming linear error propagation. That is, if we assume a different value for the uncertainty in
then the combined PDF+ uncertainty equation (28) needs to be modified as follows
everything else unchanged. It is thus clear from equation (30) that despite combined PDF sets are provided for a range of around the central value, any other choice for can be trivially implemented using the existing PDF4LHC15 sets.
Implementation in LHAPDF6. Starting from LHAPDF v6.1.6, it is possible to automatically compute the combined PDF+ uncertainties for the various cases listed above using the corresponding built-in routines19 .
PDF reweighting. Many NLO and NNLO matrix-element calculators and event generators allow the computation of PDF uncertainties without any additional CPU-time cost by means of PDF reweighting techniques. This functionality is for example available, among others, in MadGraph5_aMC@NLO [118], POWHEG [128], Sherpa [129], FEWZ [130] and RESBOS [131, 132]. These (N)NLO reweighting methods can be used in the same way for the PDF4LHC15 combined sets as for the individual sets.
In addition, an approximate LO PDF reweighting is often used, where event weights are rescaled by ratios of PDFs, see for example section 7 of the LHAPDF6 manual [115]. This LO PDF reweighting can be applied to the PDF4LHC15 combined sets, with the caveat that it is only approximately correct: (N)NLO PDF reweighting requires modifications of the LO reweighting formula. In particular, LO reweighting misses potentially large contributions from partonic channels that arise first at NLO.
Therefore, exact NLO and NNLO PDF reweighting should be used whenever possible, and the approximate LO PDF reweighting should be only used when the former is not available. The exception is of course LO event generators, where LO PDF reweighting is exact (except for the dependence of the PDF in the parton shower—but this is also true for NLO PDF reweighting).
6.3. PDF4LHC15 combined sets in the nf = 4 scheme
In addition to the combined sets listed in table 5, which are suitable for calculations in the nf = 5 scheme, PDF4LHC15 combined sets have also been made available in the nf = 4 scheme. These are required for consistent calculations where the partonic cross-sections are computed in the nf = 4 scheme, accounting for bottom quark mass effects [133].
The inputs for the nf = 4 PDF4LHC15 combination are the nf = 4 versions of the CT14, MMHT14 and NNPDF3.0 global fits. Each of these is constructed [6, 39, 134] from the corresponding nf = 5 global fits, using these as a boundary condition for from which PDFs and are obtained for using evolution equations with only four active quark flavors.
As it is well known, in the nf = 4 scheme, is rather smaller than the global average value of 0.118 for nf = 5. For example, at NLO, using nf = 4 active flavors for the running of the strong couplings up to the Z mass, one finds that corresponds to The exact value depends on the choice of mb, which is slightly different in each PDF group, though these differences are subdominant as compared to the PDF uncertainties in the combination20 . When the exact matching conditions are unknown, at an arbitrary renormalization scale μ can also be determined from the corresponding using the scheme transformation relations of [136], currently known up to four loops in QCD.
The available PDF4LHC15 combined sets in the nf = 4 scheme are summarized in table 6. Only the NLO combination is required, since no NNLO calculations in the nf = 4 scheme are yet available. Also, only Hessian sets are produced, since nf = 4 calculations are always in a region where the underlying PDF combination is essentially Gaussian.
Table 6. Same as table 5 for the combined PDF4LHC15 sets in the nf = 4 scheme. We indicate the value of in the nf = 5 scheme, the actual value in the nf = 4 scheme is substantially smaller, see text.
LHAPDF6 grid | Pert order | ErrorType | ||
---|---|---|---|---|
PDF4LHC15_nlo_nf4_100 | NLO | symmhessian | 100 | 0.118 |
PDF4LHC15_nlo_nf4_30 | NLO | symmhessian | 30 | 0.118 |
PDF4LHC15_nlo_nf4_100_pdfas | NLO | symmhessian+as | 102 | mem 0:100 → 0.118 |
mem 101 → 0.1165 | ||||
mem 102 → 0.1195 | ||||
PDF4LHC15_nlo_nf4_30_pdfas | NLO | symmhessian+as | 32 | mem 0:30 → 0.118 |
mem 31 → 0.1165 | ||||
mem 32 → 0.1195 | ||||
PDF4LHC15_nlo_nf4_asvar | NLO | — | 1 | mem 0 → 0.1165 |
mem 1 → 0.1195 |
6.4. Citation policy for the PDF4LHC recommendation
The techniques and methods presented in this report are the result of an intense collaborative efforts within the PDF4LHC community. The three reduction methods presented in section 5, the CMC-PDFs, META-PDFs and MC-H PDFs, have been substantially improved and refined (and in some cases, even developed from scratch) as a result of the fruitful discussions within the PDF4LHC working group. It is thus important to properly acknowledge this effort by providing an accurate citation policy for the usage of the PDF4LHC15 recommendations, which we spell out in some detail here:
- Whenever the PDF4LHC15 recommendations are used, this report should be cited.
- In addition, the individual PDF sets that enter the combination should also be cited:
- (1)
- (2)
- (3)
- When any of the two Hessian sets are used, PDF4LHC15_30 or PDF4LHC15_100, the original publications where the Hessian reduction methods where developed should be cited:
- (1)
- (2)MCH-PDFs [112]: S. Carrazza, S. Forte, Z. Kassabov, J. I. Latorre and J. Rojo, 'An unbiased Hessian representation for Monte Carlo PDFs', Eur. Phys. J. C 75, no. 8, 369 (2015), arXiv:1505.06736.
- When the MC sets PDF4LHC15_mcare used, the original publication where the MC compression method was developed should be cited:
- (1)CMC-PDFs [113]: S. Carrazza, J. I. Latorre, J. Rojo and G. Watt, 'A compression algorithm for the combination of PDF sets', Eur. Phys. J. C 75, no. 10, 474 (2015), arXiv:1504.06469.
- In addition, when either of the two reduction methods are employed, one should cite the original publication where the MC representation of Hessian sets was presented [101]:
- (1)G. Watt and R. S. Thorne, 'Study of MC approach to experimental uncertainty propagation with MSTW 2008 PDFs', JHEP 1208, 052 (2012), arXiv:1205.4024.
7. Future directions
We have presented in section 4 the new PDF4LHC recommendation for the computation of PDF and combined PDF+ uncertainties for LHC applications.
While the general guideline remains to combine the individual PDFs when such combination is appropriate (see section 4.1), both the way the combined uncertainties are determined, and their form of delivery, have evolved substantially since the last recommendation.
The main rationale for these changes is the observation that the PDFs of several groups have undergone significant evolution due to both new available data and methodological improvements, and that, as a consequence, there is now much better agreement between the PDF determinations based on the widest available dataset. The fact that the improvements are driven by increase of experimental information and theoretical understanding suggests that they are not accidental.
As a consequence, it now appears advisable to recommend a statistical combination of the PDFs from the three global analysis groups, with uncertainties combined with PDF uncertainties in the standard way (albeit with a conservative estimate of the uncertainty). In order to optimize and streamline usage of this statistical combination, we have produced a combined PDF4LHC15 set, both at NLO and NNLO. The set is delivered in three versions: as 100 MC replicas, or as either 30 or 100 Hessian eigenvector sets. Guidelines for usage of the PDF4LHC15 set have been presented in section 6, though it should be borne in mind that all its versions correspond to the same underlying information, and the choice of a specific version is motivated by practical considerations, such as speed versus accuracy.
Future updates of our current recommendation will require the release of new combined PDF sets, to replace the current PDF4LHC15 sets. In these future releases it might be desirable for all PDFs to use common values for heavy-quark masses, as it is currently done for the strong coupling Such future updates are likely to be motivated by two sets of considerations.
First, we expect the PDF sets which enter the current combination to undergo various updates in the coming years, as a consequence of the availability of LHC Run-II data, and also, the availability of full NNLO corrections for a variety of processes that potentially provide important constraints on PDFs and that can be accurately measured at the LHC.
Second, while only the sets included in the current combination satisfy the requirements for inclusion in the combination spelled out in section 4.1, this may change in the future. In particular, it may be possible to devise techniques for including non-global sets, using weighted statistical combination of PDF sets with uncertainties of different size or another method.
In the longer term, there are other important directions in the global PDF analysis that should be explored. Perhaps the most important one is a consistent estimation of theoretical uncertainties. While taking into account parametric uncertainties such as the value of and the heavy quark masses is feasible with the current combination methods, estimating missing higher-order uncertainties is a more challenging problem, requiring investments by the individual PDF collaborations.
Another possible future avenue is to extend the PDF4LHC combinations beyond fixed-order QCD NLO and NNLO global fits, and to perform combinations of PDF sets with improved theory such as sets with QED corrections [66, 68, 69] (needed for consistent calculations when electroweak effects are included) or of sets with threshold resummation [137] (required when soft-gluon resummation is included in the partonic cross-sections). Eventually, one might even want to further improve the accuracy of global sets by using approximate N3LO calculations when available (for example for deep-inelastic coefficient functions [138, 139]).
To summarize, we have presented a general, robust, and statistically consistent procedure for combination of PDFs. It represents a significant advancement beyond the 2010 PDF4LHC recommendation [14], and aims to meet diverse needs of the LHC Run-II programme. The advancements described here bring the modern PDFs and their combinations to a novel level of accuracy, adequate for (N)NNLO QCD computations and analysis of vast experimental information anticipated in the near future.
Acknowledgments
We are grateful to Sergey Alekhin, Johannes Blümlein, Claire Gwenlan, Max Klein, Katerina Lipka, Kristin Lohwasser, Sven Moch, Klaus Rabbertz and Reisaburo Tanaka for their feedback on this report. We are also grateful to Richard Ball, André David, Lucian Harland-Lang, Maxime Gouzevitch, Jan Kretzschmar, José Ignacio Latorre, Alan Martin, Patrick Motylinski, Ringaile Placakyte, Jon Pumplin, Alessandro Tricoli, Dan Stump, Graeme Watt, CP Yuan, as well as to many other colleagues from the PDF4LHC Working Group community for illuminating discussions about the topics presented in this report. SC and SF are supported in part by an Italian PRIN2010 grant and by a European Investment Bank EIBURS grant. SF and ZK are supported by the Executive Research Agency (REA) of the European Commission under the Grant Agreement PITN-GA-2012-316704 (HiggsTools). SF thanks Matteo Cacciari for hospitatly at LPTHE, Université Paris VI, where part of this work was done, supported by a Lagrange award. SC is also supported by the HICCUP ERC Consolidator grant (614577). The research of JG in the High Energy Physics Division at Argonne National Laboratory is supported by the US Department of Energy, High Energy Physics, Office of Science, under Contract No. DE-AC02-06CH11357. The work of PN is supported by the US Department of Energy under grant DE-SC0013681. JR is supported by an STFC Rutherford Fellowship and Grant ST/K005227/1 and ST/M003787/1, and by an European Research Council Starting Grant 'PDF4BSM'. The work of RST is supported partly by the London Centre for Terauniverse Studies (LCTS), using funding from the European Research Council via the Advanced Investigator Grant 267352. RST thanks the Science and Technology Facilities Council (STFC) for support via grant awards ST/J000515/1 and ST/L000377/1.
Footnotes
- 15
The combination can also be performed for sets with or provided that these are computed using the sets as boundary conditions for the evolution, see section 6.1.
- 16
There are arguments that a slightly higher value should be taken at NLO than at NNLO, but the majority request from PDF users is to have a common value for both orders.
- 17
Agreement between the codes for numerical DGLAP evolution at NNLO has been confirmed by a benchmarking exercise in the appendix of [111] . An already small difference between NNLO evolution of MSTW'08 grid files and HOPPET was further reduced in the MMHT14 release. At NLO, the NNPDF3.0 sets evolve according to a truncated solution of the DGLAP equation, which deviates marginally from the evolution of the CT14 and MMHT14 sets. We apply a correction to the LHAPDF grid files of the META-NLO ensemble after the combination to compensate for this weak effect.
- 18
Preliminary results on the 2015 PDG average for have been presented at https://indico.cern.ch/event/392530/contribution/1/attachments/1168353/1686119/alphas-CERN2015.pdf.
- 19
We thank Graeme Watt for the implementation of the combined PDF+ uncertainties in LHAPDF6.
- 20