“…After obtaining activations from the pre-trained networks of PICA, we built the graph by mutual k nearest neighbors between activations. Then, both the activations and the graph were used as inputs of HGCAE-P. Extensive baselines and state-of-the-art image clustering methods [35,74,19,6,3,37,43,64,50,73,75,26,71,68,9,8,67,23] were compared. Furthermore, we also trained two auto-encoder models, GAE [27], and hyperbolic auto-encoder (HAE) whose layers are hyperbolic feed-forward layers [16].…”