This paper concerns grammatical phenomena sensitive to certain classes of nominal forms, i.e., those that encode different kinds of referential properties of the nominal. We propose a grammar component for defining and picking out such semantic classes of nominal forms within typed feature structure formalisms such as the one used in HPSG, thus aiming at standardizing the representation of such phenomena. The grammar component includes four semantic features associated with the discourse referent of a nominal, i.e., cognitive status, specificity, partitivity, and whether the nominal has a universal interpretation or not. The proposed grammar component reduces to an assumed minimum a relatively large set of features that have already been proposed in analyses of the kind of phenomena at focus here, and it is hypothesized that parts of the structure are likely to be shared among grammars for different languages.
This paper summarizes ongoing efforts to provide software infrastructure (and methodology) for open-source machine translation that combines a deep semantic transfer approach with advanced stochastic models. The resulting infrastructure combines precise grammars for parsing and generation, a semantic-transfer based translation engine and stochastic controllers. We provide both a qualitative and quantitative experience report from instantiating our general architecture for Japanese-English MT using only open-source components, including HPSG-based grammars of English and Japanese.
Automatic syntactic analysis of a corpus requires detailed lexical and morphological information that cannot always be harvested from traditional dictionaries. Therefore the development of a treebank presents an opportunity to simultaneously enrich the lexicon. In building NorGramBank, we use an incremental parsebanking approach, in which a corpus is parsed and disambiguated, and after improvements to the grammar and the lexicon, reparsed. In this context we have implemented a text preprocessing interface where annotators can enter unknown words or missing lexical information either before parsing or during disambiguation. The information added to the lexicon in this way may be of great interest both to lexicographers and to other language technology efforts.
This article presents a typed feature structure grammar formalism for Mainland Scandinavian which outperforms existing grammars in several respects: it is multilingual and captures dialectal variation and many typological facts, and it is computationally efficient. Our reference point is the grammar formalism of Underwood (1997). It is proven that Underwood's formalism is intractable. Our formalism improves on this result, i.e. it is decidable in polynomial time. The article covers the phenomena covered in Underwood (1997) and non-local dependencies.
In this paper I present an incremental approach to gapping and conjunction reduction where it is assumed that the first sentence in these constructions is fully parsed before the second sentence with the elided verb is parsed. I will show that the two phenomena can be given a uniform analysis by letting the construction type of the first conjunct be carried over to the second conjunct. This construction type imposes constraints on the arguments that the second conjunct can have. The difference between gapping and conjunction reduction is captured by the already existing constructions for sentence and VP coordination. The analysis is implemented in an HPSG grammar of Norwegian.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.