2013
DOI: 10.1002/sam.11187
|View full text |Cite
|
Sign up to set email alerts
|

Understanding large text corpora via sparse machine learning

Abstract: Sparse machine learning has recently emerged as powerful tool to obtain models of high-dimensional data with high degree of interpretability, at low computational cost. The approach has been successfully used in many areas, such as signal and image processing. This paper posits that these methods can be extremely useful in the analysis of large collections of text documents, without requiring user expertise in machine learning. Our approach relies on three main ingredients: (a) multi-document text summarizatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 17 publications
(7 citation statements)
references
References 40 publications
(45 reference statements)
0
7
0
Order By: Relevance
“…While adding findings to the network is relatively easy, extracting and coding them requires more effort and time. Past research has been conducted on using machine learning to analyze accident reports [Abedin et al, 2010;Ghaoui et al, 2013;Robinson et al, 2015] so we may be able to use a similar method to teach a machine our coding scheme and easily add systems engineering failures to our solution aid.…”
Section: Discussionmentioning
confidence: 99%
“…While adding findings to the network is relatively easy, extracting and coding them requires more effort and time. Past research has been conducted on using machine learning to analyze accident reports [Abedin et al, 2010;Ghaoui et al, 2013;Robinson et al, 2015] so we may be able to use a similar method to teach a machine our coding scheme and easily add systems engineering failures to our solution aid.…”
Section: Discussionmentioning
confidence: 99%
“…We observe that IHT, in [17] is shown to solve a very similar problem to the one that we look to solve. In fact, their formulations are identical to (2), however the presentation of the algorithm's convergence and performance guarantees in [17] are restricted to a small set of matrices: those which are full column rank and meet the 3k-restricted isometry property [21].…”
Section: Algorithmsmentioning
confidence: 99%
“…Blumensath and Davies [17] present the iterative hard thresholding algorithm (IHT) as an efficient algorithm for finding a local minimum of the least-squares objective in (2). IHT has the advantage of explicitly working with the cardinality-constrained feasible set, instead of the 1 convex relaxation, controlled by λ as a proxy for the cardinality.…”
Section: Algorithmsmentioning
confidence: 99%
See 2 more Smart Citations