Controllable text generation is the primary technique for controlling specific attributes such as topic, keywords and obtaining augmented data. This work proposes a novel controllable text generation framework to improve the controllability of generation models. First, we introduce semantic control grammar, a controllable input format to generate sentences that satisfy the constraints. Second, we adopt a generation and rerank method to obtain semantically reranked controlled sentences. Extensive experiments and analyses are conducted on benchmark, natural language understanding, data-to-text generation, and text classification datasets. Through automatic evaluations, we show that our method leads to improvement over strong baselines. The results show that our model can control sentence and word attributes and semantically generate natural sentences. Furthermore, our techniques improve the generation quality of the model.
Natural language understanding (NLU) is a core technique for implementing natural user interfaces. In this study, we propose a neural network architecture to learn syntax vector representation by employing the correspondence between texts and syntactic structures. For representing the syntactic structures of sentences, we used three methods: dependency trees, phrase structure trees, and part of speech tagging. A pretrained transformer is used to propose text-to-vector and syntax-to-vector projection approaches. The texts and syntactic structures are projected onto a common vector space, and the distances between the two vectors are minimized according to the correspondence property to learn the syntax representation. We conducted massive experiments to verify the effectiveness of the proposed methodology using Korean corpora, i.e., Weather, Navi, and Rest, and English corpora, i.e., the ATIS, SNIPS, Simulated Dialogue-Movie, Simulated Dialogue-Restaurant, and NLU-Evaluation datasets. Through the experiments, we concluded that our model is quite effective in capturing a syntactic representation and the learned syntax vector representations are useful.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.