2011
DOI: 10.1007/978-3-642-19231-9_10
|View full text |Cite
|
Sign up to set email alerts
|

Patent Claim Decomposition for Improved Information Extraction

Abstract: In several application domains research in natural language processing and information extraction has spawned valuable tools that support humans in structuring, aggregating and managing large amounts of information available as text. Patent claims, although subject to a number of rigid constraints and therefore forced into foreseeable structures, are written in a language even good parsing algorithms tend to fail miserably at. This is primarily caused by long and complex sentences that are a concatenation of a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 4 publications
0
3
0
Order By: Relevance
“…While it has been argued that patent searching must become part of life science students’ information literacy instruction, 12 a more pragmatic and encompassing approach could be algorithmically mapping patents to MeSH codes, 13 which would allow integrated searches that cover both PubMed and open access patent databases. Although algorithmic parsing of typical “patent jargon” into semantics that are familiar to research scientists is only a dimly perceived possibility at this time, 14 text analysis has been able for some years to effectively identify innovative patents and provide ranking and mapping. 15 For specialized applications such as the vascular risk management literature, expert systems might provide broadly deployable semi-automated solutions within the next few years.…”
Section: Discussionmentioning
confidence: 99%
“…While it has been argued that patent searching must become part of life science students’ information literacy instruction, 12 a more pragmatic and encompassing approach could be algorithmically mapping patents to MeSH codes, 13 which would allow integrated searches that cover both PubMed and open access patent databases. Although algorithmic parsing of typical “patent jargon” into semantics that are familiar to research scientists is only a dimly perceived possibility at this time, 14 text analysis has been able for some years to effectively identify innovative patents and provide ranking and mapping. 15 For specialized applications such as the vascular risk management literature, expert systems might provide broadly deployable semi-automated solutions within the next few years.…”
Section: Discussionmentioning
confidence: 99%
“…The parser has not been optimized/retrained for the patent domain. 17 In spite of the technical difficulties (Parapatics and Dittenbach 2009) and loss of linguistic accuracy for patent texts reported in Mille and Wanner (2008), most patent processing systems that use linguistic phrases use the Stanford parser because its dependency scheme has a number of properties that are valuable for Text Mining purposes (de Marneffe and Manning 2008). The Stanford parser collapsed typed dependency model has a set of 55 different syntactic relators to capture semantically contentful relations between words.…”
Section: Stanfordmentioning
confidence: 99%
“…Due to the significant value of information in chemical patents, many research efforts have been made toward the development of more effective information extraction systems specifically for chemical patents (Parapatics and Dittenbach, 2011 ; Akhondi et al, 2014 ; Chen et al, 2020 ). Several fundamental information extraction tasks, such as named entity recognition (NER) (Zhai et al, 2019 ), and relation extraction (Peng et al, 2018 ) have been extensively investigated.…”
Section: Introductionmentioning
confidence: 99%