2006
DOI: 10.1007/11776420_45
|View full text |Cite
|
Sign up to set email alerts
|

Active Sampling for Multiple Output Identification

Abstract: We study functions with multiple output values, and use active sampling to identify an example for each of the possible output values. Our results for this setting include: (1) Efficient active sampling algorithms for simple geometric concepts, such as intervals on a line and axis parallel boxes. (2) A characterization for the case of binary output value in a transductive setting. (3) An analysis of active sampling with uniform distribution in the plane. (4) An efficient algorithm for the Boolean hypercube whe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
17
0

Year Published

2008
2008
2013
2013

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(17 citation statements)
references
References 17 publications
0
17
0
Order By: Relevance
“…An example of the underlying distribution where these assumptions are satisfied is shown in Figure 1. Note that these assumptions are much more realistic than the separable/nearseparable assumption assumed in [8] [14]. Based on our assumptions, abrupt changes in local density indicate the presence of rare classes.…”
Section: Semiparametric Density Estimation For Rarementioning
confidence: 85%
See 1 more Smart Citation
“…An example of the underlying distribution where these assumptions are satisfied is shown in Figure 1. Note that these assumptions are much more realistic than the separable/nearseparable assumption assumed in [8] [14]. Based on our assumptions, abrupt changes in local density indicate the presence of rare classes.…”
Section: Semiparametric Density Estimation For Rarementioning
confidence: 85%
“…For example, the method based on mixture models proposed in [14] is among the first attempts in this direction; in [8], the authors proposed a generic consistency algorithm, and proved upper bounds and lower bounds for this algorithm in some specific situations. Both of the two methods require that the support regions of the different classes be separable or near-separable to work well.…”
mentioning
confidence: 99%
“…However, in practice the user of a fuzzer does not usually care about the tests in a cluster, but only about finding at least one example from each set with no particular desire that it is a perfectly "representative" example. The core problem we address is therefore better considered as one of multiple output identification [8] or rare category detection [8,40], given that many faults will be found by a single test case out of thousands. This insight guides our decision to provide the first evaluation in terms of discovery curves (the most direct measure of fuzzer taming capability we know of) for this problem.…”
Section: Related Workmentioning
confidence: 99%
“…Fine et al [5] abstract the rare category detection problem as an output identification task in a learning model. The learning model has an unknown target function f which maps every input in χ to one of m output values.…”
Section: Related Workmentioning
confidence: 99%