“…We have tuned hyperparameters for DeepAffinity variants including learning rate ({10 −3 , 10 −4 }), batch size ({64, 128} (16 for CNN-GCN because of the limit of GPU memory) and dropout rate ({0.1, 0.2}) using random 10% of training data as validation sets. When HRNN was used to model protein sequences, we have also tuned k-mer lengths and group sizes in pairs [{ (40,30), (48,25), (30,40), (25,48), (15,80), (80,15)} for Davis and { (40,25), (50,20), (25,40), (20,50)} for KIBA and PDBbind] using the validation sets.…”