2021
DOI: 10.1007/978-3-030-92270-2_40
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking Binary Hyperparameters for Deep Transfer Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
9
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(11 citation statements)
references
References 22 publications
1
9
0
1
Order By: Relevance
“…Convolutional Neural Networks (CNNs) represent the predominant deep learning (DL) technique for image analysis [8], [11]. A CNN operates as a feedforward neural network designed to extract image features by applying filters, also referred to as kernels or feature detectors, to the image.…”
Section: Literature Reviewmentioning
confidence: 99%
See 1 more Smart Citation
“…Convolutional Neural Networks (CNNs) represent the predominant deep learning (DL) technique for image analysis [8], [11]. A CNN operates as a feedforward neural network designed to extract image features by applying filters, also referred to as kernels or feature detectors, to the image.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Each image was subjected to meticulous manual labeling into various predefined ship types or categories, such as submarines, cruisers, destroyers, carriers, and more. Subsequently, deep learning convolutional neural networks (CNNs) were trained to autonomously recognise these classifications based on image attributes [7], [8].…”
Section: Introductionmentioning
confidence: 99%
“…In the Deep Learning, Reinforcement Learning and Evolutionary Strategies fields, researchers are most interested in figuring out how to solve a specific set of problems and evaluate the performance of their agents/models with some loss or fitness functions. While their results are often impressive even in the generality of the found solutions (Team et al, 2021) or how prone they are to transfer learning (Liu et al, 2023;Plested and Gedeon, 2022), it is extremely difficult to afterwards take these models and "evolve them" in order to reuse as many parameters as possible and decrease the training time of a new, more complex version of the same model, or for a different task. It is this kind of complexification that we are most interested in with this line of research.…”
Section: Complexification and Open Endednessmentioning
confidence: 99%
“…Fully supervised computer vision models for problems like classification are typically trained on datasets like ImageNet (Deng et al 2009a), OpenImages (Kuznetsova et al 2018), JFT300M (Kolesnikov et al 2020;Sun et al 2017) etc., and have also proven themselves to be effective for a variety of downstream tasks via transfer learning (Plested and Gedeon 2022;Guo et al 2019;Wan et al 2019). Despite this, it is challenging to adapt these models to other domains due to various reasons including limited data and annotation overhead.…”
Section: Introductionmentioning
confidence: 99%