Seaquest data
Created by: tgiani
Implementation of seaquest data from https://arxiv.org/pdf/2103.04024.pdf. There are a number of things to be checked
- the data are not availbale from hepdata, so I copied them from the paper (please double check)
- the data are converted into data for distributions differential in hadronic rapidity and invariant mass using Eqs.(4.6), (4.7) from https://arxiv.org/pdf/1009.5691.pdf. This is done with a python script saved in the rawdata folder
- there is a single source of systematic which is considered to be fully correlated as specified in the paper. Also, my understanding is that Eq.(9) of the paper gives the correlation matrix between statistical uncertainties. This is not implemented yet. Should I include it? (I guess so but I first want to check, for the 2001 data such covariance matrix is not implemented)
- I don't understand the very last paragraph of the paper, in particular Eq.(10). This might be relevant for the computation of theory predictions with apfel
Merge request reports
Activity
requested review from @enocera
requested review from @enocera
@tgiani: I went through the paper and here are some suggestions.
- the data are not available from hepdata, so I copied them from the paper (please double check)
It seems to me that you copied the data correctly.
- the data are converted into data for distributions differential in hadronic rapidity and invariant mass using Eqs.(4.6), (4.7) from https://arxiv.org/pdf/1009.5691.pdf. This is done with a python script saved in the rawdata folder
Why do you need a python script? Cannot you perform the necessary kinematic transformations in the filter file
FTDY.cc
? In order to streamline the implementation I suggest to remove the python script and modify instead the filter file.- there is a single source of systematic which is considered to be fully correlated as specified in the paper. Also, my understanding is that Eq.(9) of the paper gives the correlation matrix between statistical uncertainties. This is not implemented yet. Should I include it? (I guess so but I first want to check, for the 2001 data such covariance matrix is not implemented)
Your understanding matches mine. Please proceed and implement the covariance matrix for statistical uncertainties. There's actually a small discrepancy between the diagonal values of the matrix in Eq.(9) and the square of the statistical uncertainties reported in Table 1, but I guess that we can live with it and use the matrix in Eq.(9) tout court.
- I don't understand the very last paragraph of the paper, in particular Eq.(10). This might be relevant for the computation of theory predictions with apfel
The thing that I don't understand is how Wu-Ki Tung, who passed away in 2009, was able to provide them with the code for the NLO computation (as they explicitly state), unless he came back from afterlife. Apart from that, I'd say that the problem concerns how to compute the theoretical predictions, not how to correct the data. I think that I can take care of that, either by implementing explicitly Eq.(10) or by interpolating the acceptance correction.
Created by: tgiani
@enocera thanks. Sure, I can perform the kinematic transformation directly in the filter if you think it s better
@enocera thanks. Sure, I can perform the kinematic transformation directly in the filter if you think it s better
@tgiani: I think it is, if you don't mind. Thanks.
Created by: tgiani
@enocera I was multiplying the data by a jacobian to convert them in data for distributions differential in rapidity, however this is wrong because these are ratios, so the jacobian simplifies between numerator and denominator. So there are no kinematic transformation to be done. The data are those in the table of the paper and the only additional thing to be computed is the hadronic rapidity. I ve removed the python script and add the covariance matrix for statistical uncertainties as well
requested review from @enocera
@Zaharid cc @tgiani The story is as follows. SeaQuest delivers a measurement of sigma(pd)/2sigma(pp) for six points as a function of xb, xt and M (all averaged), see Table 1 in 2103.04024. So far so good, this is the same observable measured by NuSea (E866) already implemented in NNPDF3.1. Unfortunately, as they say in the paper, the data in Table 1 is not corrected for detector acceptance effects. This implies, as they say towards the bottom of the second column on page 7, that the theoretical prediction must be constructed according to Eq.(10). Each measured bin is divided in 10 sub-bins; each of these bins is computed with different kinematics and receives an acceptance correction (see Extended data Table 3). After a bit of thinking, I concluded that the way to implement this kind of information is to compute 20 FK tables, 10 for the numerator of the observable and 10 for the denominator. Each of these 10 FK tables corresponds to one of the ten columns (with six rows) of Extended data Table 3. The observable then becomes a
COMPOUND
observable, should appropriate operators be defined.I've defined two operators. a)
COM
: this takes the 20 FK tables and computes Eq.(10). Note that this allowed me to implement the acceptance factors by means ofACC
K-factors (rather than to hard-code them in apfel), where each number multiplies a bin in each of the 20 FK tables. The definition of the operator has been propagated to the documentation. b)SMT
: this takes only 10 of the FK tables (either for the numerator or for the denominator) and computes their sum. This operator is required to estimate the nuclear uncertainty due to the deuteron in the numerator, as it allows one to compute the numerator with a proton-proton or with a proton-deuteron PDF.As far as I can tell the implementation is meaningful, for example see here two data/theory comparisons
- for NNPDF4.0 (at NNLO): https://vp.nnpdf.science/o8v2boMxS-qOnX5XfV8DCA==
- for CT18NLO (at NLO): https://vp.nnpdf.science/Mg8MGbyGToWDo4duh3ZoDQ== Note that, by eye, the second of these comparisons resembles Figure 1 of 2103.04024.
Finally: the measurement requires the addition of a nuclear correction for the deuteron in the numerator. This requires in turn to iterate the fit to the deuteron data with the inclusion of SeaQuest, something that I hadn't considered in the first instance. Incidentally, deuteron corrections for all of the other data sets have to be updated accordingly.
This is on its way.This has been done. Needless to say that, with hindsight, I would have strongly voiced against the decision to include SeaQuest in NNPDF4.0, given the non-negligible (and mostly unnoticed/unappreciated) work that its implementation required.As far as I'm concerned, this PR is good to go, perhaps pending successful tests and rebase on master (with updated runcards). Here's a data/theory comparison when the extra nuclear uncertainty is included (via deweighting): https://vp.nnpdf.science/RGfy_dVZQa6xYT72MFW_7w==
removed review request for @enocera