-
Notifications
You must be signed in to change notification settings - Fork 180
Adding MCH aQC for MC #1830
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding MCH aQC for MC #1830
Conversation
|
REQUEST FOR PRODUCTION RELEASES: This will add The following labels are available |
|
Hello @JianLIUhep , @aferrero2707 Can you check this one? Chiara |
|
Hi @lmassacr, is |
|
Hi @JianLIUhep, Indeed I anticipated a next PR to come by putting already the json file. I still need indeed to add the workflow for MCH+MFT and MCH+MFT+MID tracks for the MC aQC. The corresponding modifications of the dpg scripts is work in progress. My local tests are currently failing. Regarding this PR, I see that I have an error in the automatic checks (which I didn't get locally). This is related to the fact the the QC reads some intermediate .root files produced by the muon reco, for which it seems the file is eitheir corrupted or an array was not properly filled. Cheers, |
|
Hi @JianLIUhep and @aferrero2707 , I am pinging you again regarding this PR if you have suggestions on how to proceed. |
|
Hi @lmassacr, I saw some segmentation and bus errors from other tasks before the mch task crash. Maybe try to make a dummy commit to retrigger the test. |
|
@JianLIUhep @chiarazampolli @catalinristea @alcaliva @lmassacr I am getting the same kind of error locally, using O2DPG master from yesterday. The error in the CI from the full build log: I did my local test with O2DPG master, using this command (the with the sime kind of failure in |
|
Hi @aferrero2707, |
|
@sawenzel, the crash in the testing is unrelated to the code. Could you force merge this PR? |
|
@alcaliva : You are admin. You should also be able to merge. |
|
The failing test is due to a broken TPC digitization which was fixed. I would be in favour of waiting for another CI iteration to see that the development here is good. |
|
Hi @sawenzel , I just made a dummy commit to retrigger the CI. As there were some crashes in the MCH tasks in the first attempt (although maybe related to previous crash in the chain unrelated to this code), it is better to retest. Cheers, |
|
@lmassacr : Thanks... but the dummy commits are no longer necessary. I had already retriggered the CI in the github actions tab. |
|
Hi @sawenzel , |
|
Hello @sawenzel! The CI failure seems again to be completely un-related ( |
Hello @aferrero2707, I just restarted the CI. |
|
The CI was giving an error related to the alien token, however it was refreshed last week in the machine. I just restarted the test to see if it was just a case. |
|
Hi @jackal1-66, |
Pinged the machine experts |
The new failure seems to be legit, can you take a look? |
|
Hi @jackal1-66 , Indeed this one is on this PR. I have to investigate because I don't have the issue locally. The intermediate root files seem somehow corrupted. And there is a problem with the trackrofs objects. |
|
@lmassacr @jackal1-66 @sawenzel I am not 100% sure that the last error is really related to the changes in this PR. The CI fails during test 84 ( Could it be that due to that failure the intermediate ROOT files are not closed properly? Note that I have been able to reproduce the crash of |
|
@aferrero2707 I checked on the current cvmfs build and I do see the same error when checking test 84. Hence this looks definitely unrelated to this PR. |
|
Hello @jackal1-66, |
|
@lmassacr : I think the main error in the CI logs of O2fst/o2 is genuine and related to this PR. The Here is what I did: It could be that the error occurs only randomly but it looks as if the |
|
Hi @sawenzel, Thanks for your inputs. I am making locally the same tests as you. Although the workflow already crashes because of ft0fV0emcctp_digi, then tpcclusterpart1_1 , then tpcreco_2, mftDigitsQC0_local2 (see for instance attachment) After I managed to go through the reco, the QC part didn't fail (except for mftDigitsQC0_local2 in this specific example), the muon QC goes through and I have the outputs produced (see attachment) and filled with 1 track. I have attached the log of MCHTracksTaskQC_local1 and MCHTracksTaskQC_local2 |
|
@lmassacr : What platform are you on? As usual you would need to inspect the relevant log file, such as |
|
Hi @sawenzel, |
|
Hi @sawenzel, @aferrero2707, I am making now some tests on linux. I am able to reproduce the crash, though it is yet not clear to me what happens. In that case, I don't specify a json file. |
|
@lmassacr @sawenzel I think I have found the issue... it is in the command here: The The solution that worked for me was to add |
|
Thanks @aferrero2707 for the help. As I have upgraded to Sequoia, I will close this PR and open a new one with the fix, and with the additional workflows for MFT-MCH, and MFT-MCH-MID |

Adding to DPG scripts the workflows for the aQC of MCH and MCH+MID objects