Author Profiling (AP) aims at predicting specific characteristics from a group of authors by analyzing their written documents. Many research has been focused on determining suitable features for modeling writing patterns from authors. Reported results indicate that contentbased features continue to be the most relevant and discriminant features for solving this task. Thus, in this paper, we present a thorough analysis regarding the appropriateness of different distributional term representations (DTR) for the AP task. In this regard, we introduce a novel framework for supervised AP using these representations and, supported on it. We approach a comparative analysis of representations such as DOR, TCOR, SSR, and word2vec in the AP problem. We also compare the performance of the DTRs against classic approaches including popular topic-based methods. The obtained results indicate that DTRs are suitable for solving the AP task in social media domains as they achieve competitive results while providing meaningful interpretability.[2] becoming an important issue for many companies and organizations. For example, from the marketing perspective, knowing characteristics of a group of Internet users could help in improving the impact of some particular products, and, from the forensic linguistics view, knowing the linguistic profile of an author could be used as valuable additional evidence in criminal investigations.Generally speaking, the author profiling (AP) task consists in analyzing written documents to extract relevant demographic information from their authors [15], such as gender, age range, personality traits, native language, political orientation, among others. Traditionally, the AP task has been approached as a single-labeled classification problem, where the different categories (e.g., male vs. female, or teenager vs. young vs. old) stand for the target classes. The common pipeline is as follows: i) extracting textual features from the documents; ii) building the documents' representation using the extracted features, and iii) learning a classification model from the built representations. As it is possible to imagine, extracting the relevant features is a key aspect for learning the textual patterns of the different profiles. Accordingly, previous research has evaluated the importance of thematic (content-based) features [15,33] and stylistic characteristics [7]. More recently, some works have also considered learning such representations utilizing Convolutional and Recurrent Neural Networks [39,14,40].Although many textual features have been used and proposed, a common conclusion among previous research is that content-based features are the most relevant for this task. The later can be confirmed by reviewing the results from the PAN 1 competitions [35], where the best-performing systems employed content-based features for representing the documents regardless of their genre. This result is somehow intuitive since AP is not focused on distinguishing a particular author through modeling its writing style 2 ,...