2018
DOI: 10.48550/arxiv.1805.01532
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Lifted Neural Networks

Abstract: We describe a novel family of models of multilayer feedforward neural networks in which the activation functions are encoded via penalties in the training problem. Our approach is based on representing a non-decreasing activation function as the argmin of an appropriate convex optimization problem. The new framework allows for algorithms such as block-coordinate descent methods to be applied, in which each step is composed of a simple (no hidden layer) supervised learning problem that is parallelizable across … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 12 publications
(34 citation statements)
references
References 5 publications
0
34
0
Order By: Relevance
“…One of these has focused on using MIP formulations for already trained networks to provide adversarial samples that can improve network stability [1,11,23]. In a different stream, [2] propose using convex relaxations for training ANNs. The authors also explore non gradient-based approaches and initialized weights to accelerate convergence of gradient-based algorithms.…”
Section: Literature Reviewmentioning
confidence: 99%
“…One of these has focused on using MIP formulations for already trained networks to provide adversarial samples that can improve network stability [1,11,23]. In a different stream, [2] propose using convex relaxations for training ANNs. The authors also explore non gradient-based approaches and initialized weights to accelerate convergence of gradient-based algorithms.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Thus, a natural idea in implicit learning is to keep the state vector as a variable in the training problem, resulting in a higher-dimensional (or, "lifted") expression of the training problem. The idea of lifting the dimension of the training problem in (non-implicit) deep learning by introducing "state" variables has been studied in a variety of works; a non-extensive list includes [17], [3], [10], [22], [23], [6] and [15]. Lifted models are trained using block coordinate descent methods, Alternating Direction Method of Multipliers (ADMM) or iterative, non-gradient based methods.…”
Section: Related Workmentioning
confidence: 99%
“…al propose a convex optimization approach based on a low-rank relaxation using the nuclear norm regularizer [25]. In [2], Askari et al consider neural net objectives which are convex over blocks of variables. A number of recent results considered the gradient descent method on the non-convex training objective, and proved that it recovers the planted model parameters under distributional assumptions on the training data [9,22,24].…”
Section: Related Work and Contributionsmentioning
confidence: 99%