2019
DOI: 10.1609/aaai.v33i01.33016826
|View full text |Cite
|
Sign up to set email alerts
|

Weakly-Supervised Hierarchical Text Classification

Abstract: Hierarchical text classification, which aims to classify text documents into a given hierarchy, is an important task in many real-world applications. Recently, deep neural models are gaining increasing popularity for text classification due to their expressive power and minimum requirement for feature engineering. However, applying deep neural networks for hierarchical text classification remains challenging, because they heavily rely on a large amount of training data and meanwhile cannot easily determine app… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
76
0
1

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 97 publications
(77 citation statements)
references
References 3 publications
0
76
0
1
Order By: Relevance
“…c p (κ i ) is a normalization constant. Following [23], we choose the number of vMF components differently for leaf and internal categories: (1) For a leaf category C j , the number of components m is set to 1 and the mixture model degenerates to a single vMF distribution. (2) For an internal category C j , we set the number of components to be the number of C j 's children in the label hierarchy.…”
Section: Topic Modeling and Pseudo Document Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…c p (κ i ) is a normalization constant. Following [23], we choose the number of vMF components differently for leaf and internal categories: (1) For a leaf category C j , the number of components m is set to 1 and the mixture model degenerates to a single vMF distribution. (2) For an internal category C j , we set the number of components to be the number of C j 's children in the label hierarchy.…”
Section: Topic Modeling and Pseudo Document Generationmentioning
confidence: 99%
“…[26] proposes a graph-CNN based model to convert text to graph-of-words, on which the graph convolution operations are applied for feature extraction. Under weakly-supervised or dataless settings, previous approaches include HierDataless [33], WeSHClass [23] and PCNB/PCEM [41], which have been introduced in Section IV-A. All above mentioned studies focus on text data without additional information.…”
Section: Related Workmentioning
confidence: 99%
“…Recent findings demonstrate the remarkable performance of applying deep learning in motion generation [7]- [9], music processing [10], [11] and several other tasks. An advantage of using deep learning is the relatively low effort for feature engineering [12]. Deep learning also enables an end-to-end mapping between the input and output features, reducing the requirement for intermediate hand-crafted annotations [10].…”
Section: Introductionmentioning
confidence: 99%
“…Motivated by the benefits of deep learning, in this study, we explore the application of deep learning models in the generation of basic dance steps. Though deep learning offers the benefits mentioned above, applying it to dance generation can be challenging because of the following three main issues: 1) deep learning models may not generate variable-length non-linear sequences [13] such as dance; 2) the given model may not be able to constrain the motion beat [1], [2], [14] to music beat, and 3) the performance of deep learning models is proportional to the number of training datasets; thus, they require large carefully-labeled datasets for good performance [12].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation