In an effort to better understand meaning from natural language texts, we explore methods aimed at organizing lexical objects into contexts. A number of these methods for organization fall into a family defined by word ordering. Unlike demographic or spatial partitions of data, these collocation models are of special importance for their universal applicability. While we are interested here in text and have framed our treatment appropriately, our work is potentially applicable to other areas of research (e.g., speech, genomics, and mobility patterns) where one has ordered categorical data, (e.g., sounds, genes, and locations). Our approach focuses on the phrase (whether word or larger) as the primary meaning-bearing lexical unit and object of study. To do so, we employ our previously developed framework for generating word-conserving phrase-frequency data. Upon training our model with the Wiktionary-an extensive, online, collaborative, and open-source dictionary that contains over 100, 000 phrasal-definitions-we develop highly effective filters for the identification of meaningful, missing phrase-entries. With our predictions we then engage the editorial community of the Wiktionary and propose short lists of potential missing entries for definition, developing a breakthrough, lexical extraction technique, and expanding our knowledge of the defined English lexicon of phrases. . We focus on a particular aspect of Shannon's work, namely joint probability distributions between word-types (denoted w ∈ W ), and their groupings by appearance-orderings, or, contexts (denoted c ∈ C). For a word appearing in text, Shannon's model assigned context according to the word's immediate antecedent. In other words, the sequence · · · w i−1 w i · · · places this occurrence of the word-type of w i in the context of w i−1 (uniquely defined by the word-type of w i−1 ), where " " denotes "any word". This experiment was novel, and when these transition probabilities were observed, he found a method for the automated production of language that far better resembled true English text than simple adherence to relative word frequencies.Later, though still early on in the history of modern computational linguistics and natural language processing, theory caught up with Shannon's work. My guess is that phrase-adaption and generative gap-filling are very roughly equally important in language production, as measured in processing time spent on each, or in constituents arising from each. One way of making such an intuitive estimate is simply to listen to what people actually say when they speak. An independent way of gauging the importance of the phrasal lexicon is to determine its size.Since then, with the rise of computation and increasing availability of electronic text, there have been numerous extensions of Shannon's context model. These models have generally been information-theoretic applications as well, mainly used to predict word associations [4] and to extract multi-word expressions (MWEs) [5]. This latter topic has been one of extre...