-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Physics results are different in fortran and cudacpp MEs (by factors or orders of magnitude) #417
Comments
ok ok, probably this is the famous single diagram enhahncement factor?... that indeed we do not hav ein cudacpp yet |
@valassi the constant factor was what I observed initially, I think it was also on eemumu at the time |
Ok this is definitely the multichannel issue I retried with
Now I get for eemumu
And for ggtt
The difference can be up to 30% in ggtt, I guess this is the running of alphas? |
@oliviermattelaer have you already implemented multichannel in cudacpp, at least as a draft patch? Otherwise I can have a look, it should be easy. The only problem is that I see we may need "get_channel_cut", see #419. Is this something we can ignore? Thanmks |
Hi, I did implement the multi-channel in the standalone_gpu output. However, you can ignore (for the moment at least) the function get_channel_cut. |
This is a relatively old result and I think I reported it somewhere already, anyway indeed if I use a fixed renormalization scale in the runcard then the numbers ar emuch better
|
I will now slowly start integrating Stefan's alphas... |
…running alphas madgraph5#373 and madgraph5#417 All ok! Now the ratio is 1 even with running alphas ./cmadevent_cudacpp < ../../../tmad/input_app_32_NOMULTICHANNEL.txt ... RESET CUMULATIVE VARIABLE 1 3.5665107131744436 3.5664076084950689 0.99997109088191050 2 0.45750753393963106 0.45750669938725042 0.99999817587183004 3 1.0034869585255990 1.0034632350091610 0.99997635891903081 4 0.42736056050249699 0.42735994776248010 0.99999856622235761 5 0.62208668391557365 0.62207238390567887 0.99997701283395946 6 0.54427805795509532 0.54426778363100825 0.99998112302354114 7 0.42777117870194153 0.42777104754863132 0.99999969340311656 8 0.90719516691836111 0.90719169458154836 0.99999617244784877 9 0.46940856376493523 0.46940250853893922 0.99998710030778426 10 1.1211030838709168 1.1210569608644638 0.99995885926359795 11 2.0814056569772625 2.0812943357966138 0.99994651634568421 12 2.3377317419203960 2.3376751859689477 0.99997580733903979 13 0.96639301107661646 0.96636328419795126 0.99996923934845916 14 0.47219747839080861 0.47219648282845950 0.99999789163984421 15 2.9999390801816128 2.9997680469889678 0.99994298777806023 16 0.31943068952450482 0.31943233120584363 1.0000051393976617 1 0.49352008663987490 0.49351851086041021 0.99999680706113625 2 1.2387754214196172 1.2387275810481315 0.99996138091646114 3 0.32641446184474765 0.32641575510565102 1.0000039620208494 4 0.48419754170306206 0.48419268874481292 0.99998997731745587 5 1.1241654439875644 1.1241187554052836 0.99995846822855972 6 0.79551174010509906 0.79548665978611088 0.99996847272299860 7 1.1905375807398626 1.1904955772543695 0.99996471889155569 8 1.5222331121596433 1.5221598419580713 0.99995186663528290 9 20.527314687202445 20.526129514290172 0.99994226361653571 10 1.6377557144388479 1.6376744770942142 0.99995039715391154 11 0.37093260175527626 0.37093237398102596 0.99999938594167992 12 0.56844879445413321 0.56843834295622875 0.99998161400286811 13 1.5926397221564033 1.5925649447066856 0.99995304810706565 14 3.1771304607040958 3.1769466958870427 0.99994216012866766 15 0.49032034992343876 0.49031882610983851 0.99999689220812371 16 0.31941863974661144 0.31942025174076361 1.0000050466502313 Iteration 1 Mean: 0.6464E+03 Abs mean: 0.6464E+03 Fluctuation: 0.271E+03 0.645E+04 100.0%
…oper test of running alphas madgraph5#373 and madgraph5#417 All ok! Now the ratio is 1 even with running alphas ./cmadevent_cudacpp < ../../../tmad/input_app_32_NOMULTICHANNEL.txt ... RESET CUMULATIVE VARIABLE 1 3.5665107131744436 3.5664076084950689 0.99997109088191050 2 0.45750753393963106 0.45750669938725042 0.99999817587183004 3 1.0034869585255990 1.0034632350091610 0.99997635891903081 4 0.42736056050249699 0.42735994776248010 0.99999856622235761 5 0.62208668391557365 0.62207238390567887 0.99997701283395946 6 0.54427805795509532 0.54426778363100825 0.99998112302354114 7 0.42777117870194153 0.42777104754863132 0.99999969340311656 8 0.90719516691836111 0.90719169458154836 0.99999617244784877 9 0.46940856376493523 0.46940250853893922 0.99998710030778426 10 1.1211030838709168 1.1210569608644638 0.99995885926359795 11 2.0814056569772625 2.0812943357966138 0.99994651634568421 12 2.3377317419203960 2.3376751859689477 0.99997580733903979 13 0.96639301107661646 0.96636328419795126 0.99996923934845916 14 0.47219747839080861 0.47219648282845950 0.99999789163984421 15 2.9999390801816128 2.9997680469889678 0.99994298777806023 16 0.31943068952450482 0.31943233120584363 1.0000051393976617 1 0.49352008663987490 0.49351851086041021 0.99999680706113625 2 1.2387754214196172 1.2387275810481315 0.99996138091646114 3 0.32641446184474765 0.32641575510565102 1.0000039620208494 4 0.48419754170306206 0.48419268874481292 0.99998997731745587 5 1.1241654439875644 1.1241187554052836 0.99995846822855972 6 0.79551174010509906 0.79548665978611088 0.99996847272299860 7 1.1905375807398626 1.1904955772543695 0.99996471889155569 8 1.5222331121596433 1.5221598419580713 0.99995186663528290 9 20.527314687202445 20.526129514290172 0.99994226361653571 10 1.6377557144388479 1.6376744770942142 0.99995039715391154 11 0.37093260175527626 0.37093237398102596 0.99999938594167992 12 0.56844879445413321 0.56843834295622875 0.99998161400286811 13 1.5926397221564033 1.5925649447066856 0.99995304810706565 14 3.1771304607040958 3.1769466958870427 0.99994216012866766 15 0.49032034992343876 0.49031882610983851 0.99999689220812371 16 0.31941863974661144 0.31942025174076361 1.0000050466502313 Iteration 1 Mean: 0.6464E+03 Abs mean: 0.6464E+03 Fluctuation: 0.271E+03 0.645E+04 100.0%
I am about to merge PR #453 where I move back to variable renormalization scales (ie truly running alphas event by event) and all results are the same in Fortran and Cudacpp. I will close #373 about running alphas. I will keep this #417 open for the moment as I want to implement some more automatic tests that frtran and cudacpp are the same, in a less verbse way - and possibly add performance comparisons. |
@oliviermattelaer I am doing a more fine-grained comparison of Fortran and cudacpp. This is with running alphas (variable renormalization scale) but no multichannel yet. CUDA and C++ are always in excellent agreement with each other. The agreement of Fortran with them is good, but not spectacular. For ggtt there are often deviations by more than 5E-5, and what is worrying is that they are ALWAYS IN THE SAME DIRECTION... can this be a systematic bug? I am printing the min and max at the end, and I am printing indivdidual events only if deviation >5E-5 in absolute value
|
Above was double, Below is for FPTYPE=f ie float. Bigger discrepancies, but always an imbalance in direction
|
I have renamed this issue to mention differences by factors or orders of magnitude. The comparison at the permille level is now moved as a separet issue to #476 |
Ok this is a long issue about large differences between fortran and cudacpp physics results, which included
The differences at the level of orderes of magnitude or factors have now gone. There are still differences at the permille or percent level (from alphas?) that I will investigate in a separate #476 This can now be closed (by PR #465) |
I am progressing in the integration of madevent+cudacpp (#400 and beyond, PR #401 and beyond).
The main issue I am facing now is that the MEs per event are different in fortran and in cuda/cpp. I am using the same type of debugging Stefan had used in his earlier tests, just caculate OUT with Fortran and OUT2 with cuda/cpp and print out the ratio.
In gg_tt, the results were wildly different. The ratios were huge: event number (in iteration with 16), OUT, OUT2, ratio
In eemumu, the ratio is just a factor 2, constant, which is more reassuring
(By the way for eemumu I get much more than the 32 events I required, also to be debugged)
@roiser does this ring any bell?
Thanks Andrea
The text was updated successfully, but these errors were encountered: