2021
DOI: 10.1109/access.2021.3068728
|View full text |Cite
|
Sign up to set email alerts
|

Metaknowledge Extraction Based on Multi-Modal Documents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 27 publications
0
5
0
Order By: Relevance
“…Visual features can identify some regions (e.g., figures, tables), whereas textual features are critical to discriminate visually similar regions (e.g., keywords, abstract, affiliation, author names, etc.). However, single modality models have insufficient capability for layout modeling, hence multi-modal approaches have recently become more popular [9,10,37]. However, they typically contain only hundreds of labeled pages due to prohibitive labeling costs to annotate many layout objects per page, which is insufficient to train and evaluate deep learning based models [27].…”
Section: Vision-based Document Analysismentioning
confidence: 99%
See 3 more Smart Citations
“…Visual features can identify some regions (e.g., figures, tables), whereas textual features are critical to discriminate visually similar regions (e.g., keywords, abstract, affiliation, author names, etc.). However, single modality models have insufficient capability for layout modeling, hence multi-modal approaches have recently become more popular [9,10,37]. However, they typically contain only hundreds of labeled pages due to prohibitive labeling costs to annotate many layout objects per page, which is insufficient to train and evaluate deep learning based models [27].…”
Section: Vision-based Document Analysismentioning
confidence: 99%
“…Zheng et al [45] introduced four granularity levels for document modeling: documents, paragraphs, sentences, and tokens, reflecting the natural hierarchical document structure. More recently, reference [9] defined a document structure tree model to organize knowledge element extraction from documents and determine their relationships, such as juxtaposition and inclusive, between sections at different levels.…”
Section: Document Modelingmentioning
confidence: 99%
See 2 more Smart Citations
“…• • • , D 5 }.For127 each document D i , we use open source NLP models to extract the entities and relations 128 (referred to as metaknowledge semantic elements in this work) in paragraphs.129In this work, we transform the HTML script of each Wiki document web page 130 into hierarchical XML files by parsing the HTML labels, such as <h1>, <h2>, <h3>, 131 <div id="toc"...>, <p>, which represent the title, section titles, summary or para-132 graphs (referred to as metaknowledge hierarchical elements in Ref [7]…”
mentioning
confidence: 99%