2017
DOI: 10.48550/arxiv.1707.00183
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Teacher-Student Curriculum Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(42 citation statements)
references
References 0 publications
0
42
0
Order By: Relevance
“…Task weights are controlled by the ratio between a recent training loss and the loss at a previous time step, so that tasks that progress faster will be downweighted, while straggling ones will be upweighted. This approach contrasts with the curriculum learning framework proposed by Matiisen et al [14], where tasks with faster progress are preferred. Loss progress, and well as a few other signals, were also employed by Graves et al [6], which formulated curriculum learning as a multi-armed bandit problem.…”
Section: Discussion and Other Related Workmentioning
confidence: 96%
“…Task weights are controlled by the ratio between a recent training loss and the loss at a previous time step, so that tasks that progress faster will be downweighted, while straggling ones will be upweighted. This approach contrasts with the curriculum learning framework proposed by Matiisen et al [14], where tasks with faster progress are preferred. Loss progress, and well as a few other signals, were also employed by Graves et al [6], which formulated curriculum learning as a multi-armed bandit problem.…”
Section: Discussion and Other Related Workmentioning
confidence: 96%
“…The second step involves the exploration decision. We may for example select a new goal based on learning progress (Baranes and Oudeyer, 2013;Matiisen et al, 2017), selecting the goal which has shown the largest recent change in our ability to reach it. Otherwise, we can also train a generative model on the state that were of intermediate difficulty to reach, and sample a next goal from this model (Florensa et al, 2018).…”
Section: Explorationmentioning
confidence: 99%
“…Automatic curriculum generation systems often assume that some tasks cannot be tackled until other easier tasks have been solved first. Curriculum has been most explored in the supervised multi-task learning case where the agent can switch between tasks at any moment (Graves et al, 2017;Matiisen et al, 2017). Although we could use some ideas from curriculum learning in our bandit-like task, these approaches would not be compatible with more general sequential decision making tasks with delayed outcomes (e.g., a robot), which represents our ultimate goal.…”
Section: Intrinsic Rewards For Multi-prediction Learningmentioning
confidence: 99%