Hi,
I recently started using this wonderful library, but have been occasionally experiencing a small quality-of-life issue where parallel.base.ParallelSamplerBase.shutdown hangs after all the workers have finished, but the worker processes don't get joined. (I'm using the GPUSampler for basic, 1-GPU, multi-worker training.) Based on this SO post and seeing some other drain statements in the file, I think adding the following to .shutdown() patches the issue:
drain_queue(self.traj_infos_queue)
drain_queue(self.eval_traj_infos_queue)
Hope this means something to you!
Hi,
I recently started using this wonderful library, but have been occasionally experiencing a small quality-of-life issue where
parallel.base.ParallelSamplerBase.shutdownhangs after all the workers have finished, but the worker processes don't get joined. (I'm using the GPUSampler for basic, 1-GPU, multi-worker training.) Based on this SO post and seeing some other drain statements in the file, I think adding the following to.shutdown()patches the issue:Hope this means something to you!