With the sky correction (
DM-9147, DM-13388), perhaps combined with the lower detection threshold ( DM-12295), we are getting many large parent objects (> 10^6 pixels), deblending which is a burden on our compute budget. In several cases, this appears to be due to scattered light surviving the coadd process.
jbosch has proposed that we do a modest-scale background subtraction before detection, which should break up the large footprints; then we will restore the background before deblending so that the measurements have access to the original (good, large-scale) sky subtraction. There's a question about what scale to use in the background subtraction, which we can determine by running detection with different background subtraction scales, and seeing where we drop below the footprint size limits we've set in the deblender.
Jim also noted that after detection, we grow footprints by the size of the PSF in our standard detection routine. Instead of growing them all at once and allowing them to merge as we currently do, he suggests growing them individually, allowing the footprints to overlap but not merge. This might also help keep the large chains from forming.
|meas_algorithms [ 10732 ]
|obs_base [ 10719 ]
|Done [ 10000 ]
|To Do [ 10001 ]
|Done [ 10002 ]