“…While the linearization of input graph may sacrifice the inter-dependency inside input graph, some papers (Ribeiro et al, 2019(Ribeiro et al, , 2020aZhao et al, 2020) Category Output use graph encoder such as GCN (Duvenaud et al, 2015) and graph transformer Koncel-Kedziorski et al, 2019) to encode the input graphs. Others (Shen et al, 2020; try to carefully design loss functions to control the generation quality. With the development of computation resources, large scale PLMs such as GPT-2 (Radford et al, 2019), BART (Lewis et al, 2020) and T5 (Raffel et al, 2020) achieve state-ofthe-art results even with simple linearized graph input (Harkous et al, 2020;Chen et al, 2020a;Kale, 2020;Ribeiro et al, 2020b).…”