<p>We present a novel graph convolutional method called star topology convolution (STC). This method makes graph convolution more similar to conventional convolutional neural networks (CNNs) in Euclidean feature spaces. STC learns subgraphs which have a star topology rather than learning a fixed graph like most spectral methods. Due to the properties of a star topology, STC is graph-scale free (without a fixed graph size constraint). It has fewer parameters in its convolutional filter and is inductive, so it is more flexible and can be applied to large and evolving graphs. The convolutional filter is learnable and localized, similar to CNNs in Euclidean feature spaces, and maintains a good weight sharing property. To test the method, STC was compared with state-of-the-art graph convolutional methods in a supervised learning setting on nine node properties prediction benchmark datasets: Cora, Citeseer, Pubmed, PPI, Arxiv, MAG, ACM, DBLP, and IMDB. The experimental results showed that STC achieved state-of-the-art performance on all these datasets and maintained good robustness. In an essential protein identification task, STC outperformed state-of-the-art essential protein identification methods. Since the similarity between CNN and STC, some techniques of CNN can be applied in STC. We showed a study of introducing transfer learning in STC. The experimental results showed that transfer learning can be used to improve the performance of STC.<br></p>