“…Our method was implemented using the PyTorch framework, and the end-to-end training and inference both ran completely on the GPU. The generator network (9 ResNet blocks) and the discriminator network were trained from scratch for 100 epochs with early stopping by using the Adam optimizer, a learning rate of η = 0.0002 with linear decay to 0 starting at Epoch 50, momentum (0.5, 0.999), a batch size of 2, λ 0 = 1000 [19], λ BCE = 0.5, λ DICE = 0.5, λ DSAC = 0.5, and λ MLE = 0.5. For DSAC, m = 64 hypotheses are sampled based on k p = 500 points from each bounding box per patch or k p = 1300 points from each bounding box per image.…”