Details
-
Type:
Story
-
Status: Done
-
Resolution: Done
-
Fix Version/s: None
-
Component/s: ctrl_pool
-
Labels:None
-
Story Points:0.5
-
Team:Architecture
-
Urgent?:No
Description
As has been noted on Slack, the log outputs from pipeline tasks currently seem to be getting swallowed when running the pipe_drivers scripts (yes, this is gen2...) E.g. a log from a w_2021_26 run was 65M and the equivalent log on a w_2021_34 run was only 5.2M (and missing a lot of important info…)
Eli Rykoff points out:
The problem I think is that ctrl_pool wasn’t updated, and there might be some funny logging stuff here: https://github.com/lsst/ctrl_pool/blob/master/python/lsst/ctrl/pool/log.py I certainly don’t know what to do with updating the MDC.
How to reproduce:
I have attached the logs from running the following with w_2021_34:
$ singleFrameDriver.py /datasets/hsc/repo --calib /datasets/hsc/repo/CALIB --rerun private/lauren/logTest/w34 --batch-type slurm --mpiexec='-bind-to socket' --job logTest_w34 --cores 60 --time 900 --id visit=1228 ccd=0..8^10..103 |
versus running the following with w_2021_30:
$ singleFrameDriver.py /datasets/hsc/repo --calib /datasets/hsc/repo/CALIB --rerun private/lauren/logTest/w30 --batch-type slurm --mpiexec='-bind-to socket' --job logTest_w30 --cores 60 --time 900 --id visit=1228 ccd=0..8^10..103 |
The differences should be pretty obvious with a quick glance (and I'm not sure what to make of the few lines that do make it into the w34 logs that have no indication of where they originated).
Of potential note: if I run with --cores 1, the relevant log output does seem to appear.
It might be that adding these lines to that log file somewhere is all that is needed:
lgr.addHandler(lsstLog.LogHandler())
(this is what happens in command line task)