|
EEEA-N ET: ANEARLY EXITEVOLUTIONARY NEURAL |
|
ARCHITECTURE SEARCH |
|
PUBLISHED AT ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE . DOI: |
|
https://doi.org/10.1016/j.engappai.2021.104397 |
|
Chakkrit Termritthikun |
|
STEM, University of South Australia |
|
Adelaide, SA, 5095, Australia |
|
[email protected] Jamtsho |
|
College of Science and Technology |
|
Royal University of Bhutan |
|
Phuentsholing, 21101, Bhutan |
|
[email protected] |
|
Jirarat Ieamsaard |
|
Department of Electrical and Computer Engineering |
|
Faculty of Engineering, Naresuan University |
|
Phitsanulok, 65000, Thailand |
|
[email protected] Muneesawang |
|
Department of Electrical and Computer Engineering |
|
Faculty of Engineering, Naresuan University |
|
Phitsanulok, 65000, Thailand |
|
[email protected] |
|
Ivan Lee |
|
STEM, University of South Australia |
|
Adelaide, SA, 5095, Australia |
|
[email protected] |
|
ABSTRACT |
|
The goals of this research were to search for Convolutional Neural Network (CNN) architectures, |
|
suitable for an on-device processor with limited computing resources, performing at substantially |
|
lower Network Architecture Search (NAS) costs. A new algorithm entitled an Early Exit Population |
|
Initialisation (EE-PI) for Evolutionary Algorithm (EA) was developed to achieve both goals. The |
|
EE-PI reduces the total number of parameters in the search process by filtering the models with fewer |
|
parameters than the maximum threshold. It will look for a new model to replace those models with |
|
parameters more than the threshold. Thereby, reducing the number of parameters, memory usage |
|
for model storage and processing time while maintaining the same performance or accuracy. The |
|
search time was reduced to 0.52 GPU day. This is a huge and significant achievement compared to |
|
the NAS of 4 GPU days achieved using NSGA-Net, 3,150 GPU days by the AmoebaNet model, and |
|
the 2,000 GPU days by the NASNet model. As well, Early Exit Evolutionary Algorithm networks |
|
(EEEA-Nets) yield network architectures with minimal error and computational cost suitable for a |
|
given dataset as a class of network algorithms. Using EEEA-Net on CIFAR-10, CIFAR-100, and |
|
ImageNet datasets, our experiments showed that EEEA-Net achieved the lowest error rate among |
|
state-of-the-art NAS models, with 2.46% for CIFAR-10, 15.02% for CIFAR-100, and 23.8% for |
|
ImageNet dataset. Further, we implemented this image recognition architecture for other tasks, such |
|
as object detection, semantic segmentation, and keypoint detection tasks, and, in our experiments, |
|
EEEA-Net-C2 outperformed MobileNet-V3 on all of these various tasks. (The algorithm code is |
|
available at https://github.com/chakkritte/EEEA-Net ). |
|
Keywords Deep learningNeural Architecture Search Multi-Objective Evolutionary Algorithms Image classification |
|
This work is done when Chakkrit Termritthikun works as a visiting research student in University of South AustraliaarXiv:2108.06156v1 [cs.CV] 13 Aug 2021EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
1 Introduction |
|
Deep convolutional neural networks (CNNs) have been widely used in computer vision applications, including image |
|
recognition, image detection, and image segmentation. In the ImageNet Large Scale Visual Recognition Challenge |
|
(ILSVRC) Russakovsky et al. [2015], the AlexNet Krizhevsky et al. [2017], GoogLeNet Szegedy et al. [2015], ResNet |
|
He et al. [2016], and SENet Hu et al. [2018] were represented models that had been widely used in various applications. |
|
The SqueezeNet Iandola et al. [2016], MobileNets Howard et al. [2017], NUF-Net Termritthikun et al. [2019, 2020], |
|
and ShuffleNet Zhang et al. [2018] models were simultaneously developed to be used on devices with limited resources. |
|
All of these architecture networks have been strengthened and advanced by developers for many years. |
|
However, a significant drawback of the useability of these models, and to the development of the efficient CNNs models, |
|
was the dependence on the designer’s expertise and experience, including utilising resources such as high-performance |
|
computing (HPC) for the experimentation. The datasets used for analysis also affect the model efficiency, depending on |
|
the different features that different datasets manifest, and all image recognition datasets require specialised research |
|
knowledge for each dataset when modelling. One algorithm, the NAS Zoph and Le [2016] was designed to search the |
|
CNN network architecture for different datasets and thereby avoided the hitherto human intervention or design activity |
|
except during the definition of the initial hyper-parameters. |
|
The CNN network architecture is flexible, allowing the model to be developed in different structures, with the module |
|
structure consisting of layers, linked in sequence with different parameters. Models obtained from NAS methods differ |
|
in structure and parameters, making NAS searches more efficient in finding dataset models, resulting in finding a model |
|
for each particular dataset which have a unique structure and parameter set. |
|
Reinforcement Learning (RL) and gradient descent algorithms, automate the search for models of deep learning. |
|
Searching a model with an RL algorithm takes 2,000 GPU days to uncover an effective model. However, when the |
|
gradient descent algorithm is focused on searching a model with the highest accuracy for only one objective; it takes |
|
4 GPU days. Both algorithms have difficulty in dealing with a multi-objective problem. EA, however, can solve |
|
multi-objective optimisation problems. |
|
EA apply optimisation methods that mimic the evolution of living organisms in nature, including reproduction, mutation, |
|
recombination, and selection. EA can find the most suitable candidate solution with quality function rates. EA-based |
|
NAS approaches are very robust with shallow error values for experiments with CIFAR-10 and CIFAR-100 datasets. |
|
However, past search algorithms in models of the EA-based NAS approaches have taken up to 3,150 GPU days in Real |
|
et al. [2019], 300 GPU days in Liu et al. [2018a], and 4 GPU days in Liu et al. [2018b]. Many network architectures |
|
built from NAS have a high number of parameters and high computing costs. It is obvious that extensive network |
|
architectures must be avoided when attempting to identify a network architecture that is suitable for other applications. |
|
A network architecture, built from the DARTS Liu et al. [2018b] search space, is a multi-path NAS. However, excessive |
|
path-level selection is a problem for multi-path NASs where every operation of the super network of a multi-path NAS |
|
takes a separate path. This means that all connected weights across all paths require considerable memory. |
|
The single-path NAS Stamoulis et al. [2019] was introduced to solve the NAS problem by finding a subset of kernel |
|
weights in the single layer. This layer is called a super kernel. We called the model built from the super kernel, the |
|
Supernet. Based on the Supernet, which has many subnets, whole subnets are trained at the same time through weight |
|
sharing. The Supernet can be sampled and deployed without re-training. |
|
In our current work, we developed the EE-PI method for Evolutionary NASs. EE-PI was applied to the Multi-Objective |
|
Evolutionary Algorithm (MOEA) in the first generation to locate the newly created model with a certain number of |
|
parameters, discarding models with parameters more than the set criteria. This process iteratively continues until the |
|
model with fewer parameters than the initial set criteria is found. Thus, EE-PI complements MOEA. We created this |
|
add-on to prevent models with a high number of parameters and a high number of Floating point Operations Per Second |
|
(FLOPS). Also, using a small number of parameters helps reduce model search costs. |
|
The key contributions of this paper are: |
|
•We introduce Evolutionary Neural Architecture Search (EA-Net), which adopts a multi-objective evolutionary |
|
algorithm for neural architecture search with three objectives: minimisation of errors, minimal number of |
|
parameters, and lowest computing costs. |
|
•We proposed a simple method called Early Exit Population Initialisation (EE-PI) to avoid a model with a high |
|
number of parameters and high computational cost, by filtering the neural network architecture based on the |
|
number of parameters in the first-generation of the evolution. The architectures obtained by this method are |
|
called Early Exit Evolutionary Algorithm networks (EEEA-Net). |
|
2EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Image |
|
Softmax x N |
|
x N |
|
x N Normal |
|
Normal Normal |
|
Normal X1 X2 Image |
|
Initial channel |
|
X1 |
|
X1 |
|
X1 X2 |
|
X2 |
|
X2 Channel |
|
increment |
|
Reduction |
|
Normal Normal |
|
Reduction Normal |
|
Figure 1: NASNet Zoph and Le [2016] network architecture (left) and NSGA-Net Lu et al. [2019] network architecture |
|
(right). |
|
•We conduct extensive experiments to evaluate the effectiveness of an EEEA-Net by outperforming MobileNet- |
|
V3 for all of the image recognition, object detection, semantic segmentation, and keypoint detection. Also, |
|
the EEEA-Net was widely tested on standard CIFAR-10 Krizhevsky [2009], CIFAR-100 Krizhevsky [2009], |
|
ImageNet Russakovsky et al. [2015], PASCAL VOC Everingham et al. [2010], Cityscapes Cordts et al. [2016], |
|
and MS COCO Lin et al. [2014] datasets. |
|
2 Related Work |
|
NAS was designed to search and design the model structure that suits the best to the applied dataset. Thus, the model |
|
obtained by NAS has a small number of parameters with high performance. NAS can find models suitable for both |
|
small and large datasets. The NAS can be of single-objective NAS and multi-objective NAS: a single-objective NAS |
|
considers models from a single objective such as error rate, number of parameters, or FLOPS. The multi-objective NAS |
|
considers models considering more than one objective, which we have adopted in this paper to optimise the model |
|
performance. |
|
2.1 Network Architecture Search |
|
Setting optimised parameters in each layer, such as kernel size, kernel scroll position (stride), zero paddings, as well as |
|
the output size, is the main challenge in creating CNN architectures efficiently for a given dataset. The total parameters |
|
are directly proportional to the number of layers. Manually designing a model takes too long and requires considerable |
|
experimentation to achieve optimal performance, which is why an automated model discovery is essential. |
|
The ability of NASs to find automated models suitable for datasets has proved a popular area of experimentation. Deep |
|
learning also is gaining popularity and is now being widely used. The NAS model has also been designed and expanded |
|
to enable applications in new tasks such as a NAS for semantic image segmentation, NAS for object detection, and |
|
NAS for skin lesion classification. The approaches used in NAS are RL Zoph and Le [2016], EA Real et al. [2019], Liu |
|
et al. [2018a], Lu et al. [2019], and relaxation Liu et al. [2018b]. The three critical steps for NAS are: |
|
Step 1: Search the CNN model’s search space to find the suitable architecture. A CNN architecture search model |
|
contains many search spaces which are optimised in the image classification application. Google Brain has launched a |
|
search space for the NASNet model. It is a feed-forward network in which layers are broken into subgroups called cells. |
|
The normal cell can learn from images, and the normal cell maintains an image output equal to the input size. However, |
|
the kernel stride in each reduction cell is 2, halving the input image size. Normal and reduction cells are linked. Each |
|
cell is stacked during modelling, where Nnormal cells are connected. Reduction cells are added between the Nnormal |
|
cells, as shown in Fig. 1 (left), to halve the image size, helping the next normal cell to process faster. |
|
3EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
The output from the search space thereby includes normal cells and reduction cells used in the evaluation. NAS has |
|
directed acyclic graphs (DAGs) connection between input X1andX2of cells, as shown in Fig. 1 (right). In each |
|
normal cell, there are two input activations and one output activation. In the first normal cell, input X1andX2are |
|
copied from the input (image). The next normal cell uses, as input, both the X1from the last normal cell, and the X2 |
|
from the second to last normal cell. All cells are connected the same way until the end of the model. Also, each cell has |
|
a greater number of cell output channels in each layer. |
|
Step 2: Evaluate the CNN model on a standard dataset for benchmarks. These benchmarks include number of errors, |
|
number of parameters, and search cost. The normal cells and reduction cells that are found in the search space are |
|
evaluated to measure the error rate, the number of parameters, and the computing costs, using CIFAR-10. Due to |
|
limited processor resources and GPU memory, parameters such as cell count ( N), number of epochs, initial channel, |
|
and channel increment, are different for each search space and evaluation. |
|
Step 3: Evaluate with a large-scale dataset. When the model from the search space has been identified, it is evaluated |
|
with a larger dataset. Model evaluation with the CIFAR-10 dataset cannot be compared with other models because the |
|
CIFAR-10 dataset contains only ten classes. Given this constraint, CIFAR-100 datasets with 100 classes are required. |
|
The search space of NASNet used RL and was tested with the CIFAR-10 dataset, which takes up to 2,000 GPU days to |
|
model. The AmoebaNet Real et al. [2019], based on an evolutionary algorithm model, takes up to 3,150 GPU days for |
|
the same dataset. Also, the search space of NASNet was designed to use shorter search times. However, the sequential |
|
model-based optimisation (SMBO) method Liu et al. [2018c] takes 335 GPU days, the gradient descent method Liu |
|
et al. [2018b] takes just 4 GPU days, whereas weight-sharing across different structures Pham et al. [2018] takes only |
|
0.5 GPU days. |
|
As indicated, the AmoebaNet takes 3,150 GPU days, whereas the NSGA-Net Lu et al. [2019], which uses a multi- |
|
objective evolutionary algorithm to find models, takes 4 GPU days. However, although the error rate of NSGA-Net is |
|
higher than that of AmoebaNet, based on a standard CIFAR-10 evaluation, the main focus of this area of research has |
|
been the reduction of search costs. |
|
2.2 Multi-objective Network Architecture Search |
|
NAS aims to minimise errors, hyper-parameters, FLOPS, and delays, making it challenging to identify a network |
|
architecture suitable for each objective simultaneously. Thus, the best network architecture should reduce or minimise |
|
all of these dimensions. For the evolution-based NASs, NSGA-Nets Lu et al. [2019] considers FLOPS and error count, |
|
CARS Yang et al. [2020] and LEMONADE Elsken et al. [2018] consider device-agnostic and device-aware objectives. |
|
In our work, however, we sought the achievement of the three goals; minimising errors, and reducing parameters and |
|
FLOPS, simultaneously. |
|
The NASs mostly focus on creating a network architecture for image recognition, then transferring that architecture to |
|
other tasks. However, for object detection and semantic segmentation, the same network architecture can be used as a |
|
backbone. |
|
Many network architectures, such as the EfficientNet Tan and Le [2019], FBNetV2 Wan et al. [2020], DARTS Liu et al. |
|
[2018b], P-DARTS Chen et al. [2019], CDARTS Yu and Peng [2020], CARS Yang et al. [2020], LEMONADE Elsken |
|
et al. [2018], NSGA-Net Lu et al. [2019] and NSGA-NetV2 Lu et al. [2020a] were tested only on image recognition |
|
datasets. It is challenging to design and evaluate a network architecture for general purposes. |
|
Table 1 shows the research objectives of the various NASs, illustrating that the image identification architectures were, |
|
in some cases, transferred to object detection, with one, MobileNetV3 Howard et al. [2019] also being applied to |
|
transfer specifically researched image identification architecture to both object detection and semantic segmentation. |
|
Our objective was to extend this image identification architecture, using the ImageNet dataset, to object detection, |
|
semantic segmentation, as well the further purpose of keypoint detection. |
|
2.3 Inverted Residuals Network (IRN) |
|
The Inverted Residuals Network (IRN) Tan et al. [2019] concept is needed to reduce the Residuals Network (RN) |
|
parameters. In contrast, the RN concept integrates data from the previous layer into the last layer. Fig. 2 (left) shows |
|
that the RN structure has three layers: wide, narrow, and wide approach layers. The wide layers have N16output |
|
channels whereas the narrow layers have N16channels each. The wide approach layer has N32output channels |
|
(Nis the input channels in each case). However, all the convolution layers used the standard convolution. Batch |
|
normalisation (BN) and activation functions (ReLU) were also added into each convolution layer. |
|
4EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Methods Search Method Multiple Objective Dataset Searched Architecture transfer y |
|
MobileNetV3 Howard et al. [2019] RL + expert - ImageNet IR, OD, SS |
|
EfficientNet Tan and Le [2019] RL + scaling - ImageNet IR |
|
FBNetV2 Wan et al. [2020] gradient - ImageNet IR |
|
DARTS Liu et al. [2018b] gradient - CIFAR-10 IR |
|
P-DARTS Chen et al. [2019] gradient - CIFAR-10, CIFAR-100 IR |
|
PC-DARTS Xu et al. [2020] gradient - CIFAR-10, ImageNet IR, OD |
|
CDARTS Yu and Peng [2020] gradient - CIFAR-10, ImageNet IR |
|
CARS Yang et al. [2020] EA Yes CIFAR-10 IR |
|
LEMONADE Elsken et al. [2018] EA Yes CIFAR-10, CIFAR-100, ImageNet64 IR |
|
NSGA-Net Lu et al. [2019] EA Yes CIFAR-10 IR |
|
NSGA-NetV2 Lu et al. [2020a] EA Yes ImageNet IR |
|
EEEA-Net (this paper) EA+ EE-PI Yes CIFAR-10, ImageNet IR, OD, SS, KD |
|
yIR = Image Recognition, OD = Object Detection, SS = Semantic Segmentation, KD = Keypoint Detection. |
|
Table 1: Comparison of different NAS search method with multi-objectives. |
|
Inverted Residuals Network |
|
Input narrow |
|
Add Conv 1x1, f=N |
|
BN narrow |
|
approach wide |
|
Depthwise 3x3 |
|
BN |
|
ReLU Conv 1x1, f=Nx16 |
|
BN |
|
ReLU channels = N |
|
channels = Nx16 |
|
channels = Nx16 |
|
channels = N Xi |
|
Xi+1 Residual Network |
|
Input wide |
|
Add Conv 1x1, f=Nx32 |
|
BN |
|
ReLU wide |
|
approach narrow |
|
Conv 3x3, f=Nx16 |
|
BN |
|
ReLU Conv 1x1, f=Nx16 |
|
BN |
|
ReLU channels = N |
|
channels = Nx16 |
|
channels = Nx16 |
|
channels = Nx32 Xi |
|
Xi+1 |
|
Figure 2: The difference between the residual network (left) and the inverted residual network (right). |
|
The RN structure is modified and reversed to obtain the IRN. The layers in IRN are defined as narrow layer, wide layer |
|
and narrow approach layer. In IRN, the number of output channels obtained is equal to the number of input channels, |
|
N, as shown in Fig. 2 (right). When the data is fed into the 11convolution layer, the number of channels will be |
|
expanded to N16. The wide layer changes to a 33depth-wise separable convolution instead of a 33standard |
|
convolution, reducing the FLOPS and number of parameters. There are N16channels in a wide layer, which is equal |
|
to the previous layer. Also, 11standard convolution is used in the narrow approach to reduce the channels’ size to be |
|
equal to the input channel N. Then, the input data ( Xi) is combined with the IRN’s output to get data ( Xi+ 1). |
|
Convolutional layers can be defined in various formats to find a model structure with EA. Convolutional layers are also |
|
called cells. The types of convolutional layers used in our experiment are presented in Table 2. |
|
3 Method |
|
The most commonly used methods to develop NAS are RL and gradient descent algorithms. However, these algorithms |
|
possess limitations in solving multi-objective problems. EA automates the model search process, is easier to implement, |
|
and enables discovery of solutions while considering multiple objectives. |
|
A general description of an EA, including encoding, a presentation of the multi-objective genetic algorithm, and genetic |
|
operations used with NAS, is provided in Section 3.1. The Early Exit Population Initialisation concept and method, its |
|
simple, yet effective application in EA, mitigating the complexity and parameters used in the previous models, while |
|
maintaining the same accuracy, are described in Section 3.2. |
|
3.1 Evolutionary Neural Architecture Search |
|
The Genetic Algorithm (GA) is an algorithm based on Darwinian concepts of evolution. The GA is part of a random- |
|
based EA Xie and Yuille [2017], Baldominos et al. [2017], Real et al. [2017]. GA’s search-based solutions stem from |
|
the genetic selection of robust members that can survive. The population is integral to GA because GA solutions are |
|
5EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
kernel Type Layer |
|
3×3 max and average pooling |
|
3×3 and 5×5 depth-wise separable convolution |
|
3×3 and 5×5 dilated convolution |
|
3×3 and 5×5 inverted residuals convolution |
|
- skip connection - |
|
Table 2: Search space of EA-Net and EEEA-Net. |
|
inv |
|
3×3 |
|
inv |
|
5×5+avg |
|
3×3 |
|
max |
|
3×3Input 1 |
|
( X1) |
|
Input 2 |
|
( X2)dep |
|
3×3 |
|
skip |
|
Concat |
|
max |
|
3×3 |
|
dep |
|
5×5+ |
|
++ |
|
Output 0 |
|
12 |
|
3L = Type of conv layers [0,1,2,3,4,5,6,7,8] |
|
A = Index of cells [ 0] |
|
B = Index of cells [ 0,1] |
|
C = Index of cells [ 0,1,2] |
|
D = index of cells [ 0,1,2,3]Input-1 |
|
Input-2 Output |
|
102 |
|
383 |
|
01 |
|
76 |
|
20 |
|
Chromosome = LA1LA2,LB1LB2,LC1LC2,LD1LD2 |
|
|
|
= 3080-0121-1202-6373 |
|
Figure 3: The chromosome structure of the Evolutionary Algorithm. |
|
like organisms that evolve after the environment. The most suitable solutions need to rely on genetic diversity. Thus, a |
|
greater number of genetically diverse populations enables more effective GA solutions. |
|
The initial phase of GA creates the first population for candidate solutions. The population is determined by population |
|
size, which describes total solutions. Each solution is called an individual, where an individual consists of a chromosome. |
|
The chromosome is a mix of genes. In the initial population, it is possible to provide unique information about each |
|
gene to all other genes, at random. A fitness function computes the fitness value for each individual. CNN’s model |
|
structure is searched with the NAS search space, defining error rate as a fitness function where fitness value represents |
|
dataset error value. As shown in Equation 1, the fitness value is calculated where n is the number of individuals. |
|
fitness (i) =f(xi);i= 1;2;3;::;n (1) |
|
Organisms consist of different phenotypes and genotypes. Appearances such as foreign features (such as eye colour) |
|
and internal features (such as blood types) are called phenotypes. Genes of different organisms, called genotypes, can |
|
be transferred from model to model by gene transfer. |
|
The CNN model’s architecture is represented as a genotype in the NAS search space. A subset of the NAS search space |
|
includes normal cells and reduction cells. The cells are stacked in the complete architecture. Normal cells or reduction |
|
cells consist of connected layers such as convolution layers, average pooling, max pooling, and skip connection. A |
|
complete model must connect cells to create a genotype for training or testing. |
|
3.1.1 Encoding |
|
The genotype of the NAS model consists of normal cells and reduction cells, called chromosomes. There are various |
|
genes linked to chromosomes. The number of genes is defined as LA 1LA 2,LB 1LB 2,LC 1LC 2, andLD 1LD 2as |
|
in Equation 2. The gene consists of operations ( L) and indices of operations ( A;B;C;D ). Operation ( L) can be |
|
considered a type of CNN layer, such as max pooling, average pooling, depth-wise separable convolution, dilated |
|
convolution, inverted residuals block, and skip connection. |
|
chromosome (x) =LA 1LA 2;LB 1LB 2;LC 1LC 2;LD 1LD 2 (2) |
|
For example, consider nine different operations ( L) in the experiment, as [0;1;2;3;4;5;6;7;8]. Moreover, the operation |
|
index (A;B;C;D ) refers to the operation location to be connected with other operations ( L). From Fig.3 (left), the |
|
6EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
index is defined as follows: A = [0], B = [0;1], C = [0;1;2], and D = [0;1;2;3]. The connection between operations (L) |
|
and the operation index ( A;B;C;D ) determines the location of the connection between operations. For example, LA’s |
|
gene code,LA 1LA 2is 30,80, meaning output data processed in operation 3 and 8 will be linked to index 0. |
|
Similarly,LB 1LB 2equals 01,21, meaning data processed by operation 0 and 1 are connected at index 1. However, in |
|
the genes of LC 1LC 2andLD 1LD 2, those genes are linked sequentially to the output of LA 1LA 2andLB 1LB 2, to |
|
help reduce the number of model parameters. If the LC 1LC 2andLD 1LD 2is connected to the same input as LA 1LA 2, |
|
andLB 1LB 2(parallel network) will increase the processing time and the parameters. |
|
PreviousIndex =index 2 (3) |
|
The position of the previous index can be computed from Equation 3, where the index is greater than 1. If the index |
|
is an even number, it is linked to the even previous index; otherwise, it is linked to the odd previous index. Thus, the |
|
LC 1LC 2gene is 12-02, which has an index of 2, and the LC 1LC 2gene is linked from index 0. While the LD 1LD 2 |
|
gene is 63-73, it has an index of 3. The LD 1LD 2gene is linked from index 1. However, if there are different indices in |
|
a gene, for example, a gene 63-72, operator 6 is connected from index 1 and operator 7 is from index 0. |
|
3.1.2 Multi-objective Genetic Algorithm |
|
Initially, GA was used for single-objective optimisation problems (SOOP) and, later, GA was developed to solve |
|
the multi-objective optimisation problem (MOOP) Deb et al. [2002], which has more than one objective function to |
|
minimise fitness values. The GA that can solve the MOOP problems Carrau et al. [2017], Hasan et al. [2019] is called a |
|
multi-objective genetic algorithm (MOGA). |
|
minff1(x);f2(x);:::;f k(x)g |
|
s.t.x2X(4) |
|
The optimisation of the CNN model is generally a problem with more than one objective. As illustrated in Equation 4, |
|
wherefis fitness values, the integer k2is the number of objectives, xis individual, and Xis the set of individuals. |
|
All these objectives must be as small as possible. |
|
Indicators used to measure CNNs model performance include model accuracy, model size, and processing speed. There |
|
are three objectives to consider during a model search: lowest validation error, minimum parameters, and computational |
|
cost. |
|
minfError (x);FLOPS (x);Params (x)g |
|
s.t.x2X |
|
werror +wflops +wparams = 1 |
|
werror;wflops;wparams>= 0(5) |
|
The evolutionary algorithm finds the most effective model for each objective by finding the lowest objective values of |
|
the entire population. We defined the three objective values as being equally important. Thus, it is necessary to set the |
|
weight of each of the three objective values to 1/3 to find the best model for each value. As illustrated in Equation 5, |
|
wherexis individual, Xis the individual’s set and werror;wflops;wparams weighs each objective’s weight. |
|
For the MOOP problem, it is almost impossible to find one solution that provides the optimal value for each objective |
|
function. For each solution given by the MOOP, the best solution group is called nondominated or Pareto optimal |
|
because these solutions are compared using the Pareto Domination principle. Many solutions can be obtained from the |
|
search using MOOP. These solutions will be reviewed again to find the best solution within the searched solution group. |
|
The best solution group should not dominate when compared to other solutions. For example, any solution vover- |
|
whelming a better solution can be represented as vw. If no solution vis worse than solution w, then solutions v is |
|
better thanw. |
|
3.1.3 Genetic Operations |
|
The processes used to create offspring in the new generation are called genetic operations. A new population must |
|
replace an ancestor group that cannot survive. The population can be created in two ways, by crossover or mutation. |
|
7EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Parent 2 Child = |
|
: Common |
|
: Normal |
|
: Mutant |
|
(Parent 1) : 40-30- 61-31-00-60-42-13 |
|
(Parent 2) : 40-30- 21-61-22-72-53-11 |
|
(Child) : 40-30- 61-61-02-72-52-13 |
|
(Original) : 40-30-61-61-02- 72-52-13 |
|
(Mutant) : 40-30-61-61-02- 33-52-13 |
|
Crossover : Parent 1 ⊗: Common |
|
: Parent 1 |
|
: Parent 2 |
|
Mutation : Normal Mutant |
|
Figure 4: Crossover operation (top): the parent has two different network architectures; the chromosome of each parent |
|
architecture can be visualised as a digits string. Child architecture was mixed with a chromosome index from their |
|
parent’s chromosome. Mutation operation (bottom): The location of a normal chromosome was randomly selected. |
|
Then this pair of genes will be replaced by a new random pair. |
|
Crossover is the creation of a new population by switching genes of different chromosomes from two populations. The |
|
genotype of the parent chromosomes will be recombined to create a novel chromosome, which can be done in various |
|
ways. For example, a point crossover that performs random cutting points or chromosomes to produce offspring is a |
|
crossover between two chromosomes with a random probability of 0.5. Crossover creates offspring using random genes |
|
from parents. |
|
Fig. 4 (top) demonstrates the uniform crossover operation used in this implementation, requiring two-parent architectures. |
|
We visualised the architectures as follows: 40-30-61-31-00-60-42-13, as parent 1 and 40-30-21-61-22-72-53-11, as |
|
parent 2. Then, in the crossover operation, the random probability of 0.5 is defined. The fifty-fifty chance was used |
|
to cross the gene between the two parent architectures for child modelling (40-30-61-61-02-72-52-13). The common |
|
parent gene is coloured black, but if the gene derived from the first parent is red, then the gene derived from the second |
|
parent is represented by blue. |
|
A mutation is an operation to reduce population uniformity and contribute to genetic diversity. The mutation changes |
|
data in the gene by randomly locating the gene and replacing the original gene with random new genes. The mutation |
|
causes offspring chromosomes to be different from parents. The individual being mutated is called a mutant. |
|
Fig. 4 (bottom) shows the mutation operation used during implementation; the location of a chromosome of architecture |
|
was determined by randomly selecting only one pair of gene locations (72, orange). Then it was replaced with a random |
|
pair of gene value (33, magenta) of the newly mutated gene. |
|
Begin Generate the |
|
initial populations Encoding Evaluate |
|
populations Stop ? Multi-Objective |
|
Selection Crossover |
|
No |
|
YesMutation |
|
Output |
|
FLOPsParamsError |
|
MinimizeMinimize |
|
Minimize |
|
Non-dominated |
|
Population Define maximum |
|
of parameters ( 𝞫)searching |
|
model? Param |
|
< 𝞫 ?No |
|
return the model |
|
with the less |
|
parameter Yes |
|
Early Exit |
|
Figure 5: An Early Exit Evolutionary Algorithm. |
|
8EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Algorithm 1 Multi-objective evolutionary algorithm with an Early Exit Population Initialisation. |
|
Input: The number of generations G, population size n, validation dataset D, objectivesw. |
|
Output: A set ofKindividuals on the Pareto front. |
|
Initialisation: An Early Exit Population Initialisation P1andQ1. |
|
fori= 1toGdo |
|
Ri=Pi[Qi |
|
for allp2Rido |
|
Train model ponD |
|
Evaluate model ponD |
|
end for |
|
M=tournament-selection (Pi+1) |
|
F=non-dominated-sorting (Ri) |
|
Picknindividuals to form Pi+1by ranks and the crowding distance weighted bywbased on Equation 5 |
|
Qi+1=crossover (M)[mutation (M) |
|
end for |
|
SelectKmodels at an equal distance near Pareto front from PG |
|
3.2 Early Exit Population Initialisation (EE-PI) |
|
EA can find cell patterns by selecting the best model with the lowest error rate. However, the discovery process takes |
|
longer to search and select cells in each generation. Each population has to be trained and evaluated, which increases |
|
the time needed to find cell patterns. The single-objective EA uses the error value to find network architecture. However, |
|
the network architecture with the lowest error may have too many parameters. In our experiment, the maximum number |
|
of generations was set to 30 with 40 populations per generation due to the limited resources of a single GPU. |
|
The population is the only factor affecting processing time. The evaluation must examine the entire population to select |
|
the population with the lowest error rate, thus obtaining an effective model structure. However, a longer search time is |
|
required to evaluate every population using a single processor unit. Therefore, the EE-PI method was introduced into |
|
the evolutionary algorithm to reduce the search time and control the number of parameters, as illustrated in Fig. 5 and |
|
detailed in Algorithm 1. |
|
The EE-PI method filters the CNN models based on the number of parameters in the network architecture, which is |
|
iteratively compared to a pre-set maximum value ( ). The EE-PI obtains the CNN network architecture which has less |
|
parameters than the maximum number of parameters attached to the EA, as illustrated in Fig. 5, which shows the Early |
|
Exit as the dashed-line block. |
|
EarlyExit (;), |
|
1;if |
|
0;otherwise(6) |
|
In Equation 6, where is the parameter of the model which is discovered, is the maximum number of parameters. If |
|
the number of parameters found in the model are less than or equal to the maximum number of parameters ( ), |
|
then the model will be considered as a part of the first-generation population. |
|
For example, to select a network architecture with a maximum of 3 million parameters ( = 3), EA selects the model |
|
by considering the number of parameters lower than the maximum number of parameters. Suppose network architecture |
|
is not considered, because it has more than maximum parameters ( ). In this case, it chooses a new structure with less |
|
than 3 million parameters. Therefore, in conjunction with the EA in the selection process, Early Exit facilitates the |
|
filtering out of the model with the number of parameters greater than the maximum number of parameters. The best |
|
network architecture is also discovered using the EA with Early Exit by considering the error rate and the number of |
|
parameters. |
|
4 Experiments and Results |
|
Experiments were carried out in three parts: First, finding and evaluating the network architecture with EEEA on |
|
CIFAR-10 and CIFAR-100. The second part was the finding and evaluation of the EEEA-Net using the ImageNet |
|
datasets. In the third part, the EEEA-Net obtained from the second part is applied for other tasks such as object detection, |
|
semantic segmentation, and keypoint detection. The PyTorch deep learning library was used in the experiments. The |
|
9EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
experiment was carried out on Intel(R) Xeon(R) W-3235 CPU @ 3.30GHz 12 Core CPU, 192 GB RAM and NVIDIA |
|
RTX 2080 Ti GPU, running on the Ubuntu 18.04.3 operating systems. |
|
4.1 CIFAR-10 and CIFAR-100 datasets |
|
This subsection searched for a model with the CIFAR-10 dataset; it was evaluated on CIFAR-10 and CIFAR-100 |
|
datasets. Both CIFAR-10 and CIFAR-100 datasets consisted of 60,000 images, with 50,000 images and 10,000 images |
|
in the training set and test set, respectively. The CIFAR-10 and CIFAR-100 have 10 and 100 classes, respectively, with |
|
600 images in each class. |
|
4.1.1 Architecture Search on CIFAR-10 |
|
Thirty generations with 40 populations in each generation were defined to locate the network architecture with EA. The |
|
first-generation populations were randomly generated, with subsequent populations in generations 2-30 being evolved |
|
with EA. Each population was defined with a depth of two normal cells instead of the usual six normal cells. Thus, it |
|
reduced the search time of the network architecture. The search and evolution process happens more rapidly when |
|
Early Exit is used in the initial populations’ process. Early Exit selects the network architecture having less than the |
|
pre-specified maximum of parameters ( ). Thus, population evolution will choose only network architectures that are |
|
efficient and have fewer parameters. |
|
The hyper-parameters for the search process were defined as: the total number of cells (normal cells and reduce cells) |
|
equal to eight layers with 32 initial channels by training the network from scratch for one epoch on the CIFAR-10 |
|
dataset. The hyper-parameters used included a batch size of 128, with SGD optimiser with weight decay equal to |
|
0.0003 and momentum equal to 0.9. The initial learning rate was 0.05. Using the cosine rule scheduler, the Cutout |
|
regularisation had a length set to 16, a drop-path of the probability of 0.2, and the maximum number of parameters |
|
equal to 3, 4, and 5 million. |
|
The evolutionary algorithm (EA-Net, = 0) took 0.57 GPU days to find the network architecture with NVIDIA RTX |
|
2080 Ti. However, the early exit evolutionary algorithm (EEEA-Net-A, = 3) took 0.38 GPU days, EEEA-Net-B |
|
(= 4) took up to 0.36 GPU days, and EEEA-Net-C ( = 5) took up to 0.52 GPU days. These architectures are used |
|
for performance evaluation in the next section. |
|
4.1.2 Architecture Evaluation on the CIFAR-10 dataset |
|
The network architecture had to be changed to find the normal and reduced cells with the Early Exit evolutionary |
|
algorithms. The CIFAR-10 dataset was used for the evaluation. The hyper-parameters were defined with the number of |
|
all cells (normal and reduce cells) set to 20 layers with 32 initial channels, the network was trained from scratch with |
|
600 epochs with a batch size of 96, SGD optimiser with weight decay was 0.0003 and momentum 0.9, and the initial |
|
learning rate set to 0.025. Using the cosine rule scheduler, the Cutout regularisation had a length set to 16, a drop-path |
|
of the probability of 0.2, and auxiliary towers of weight equal to 0.4. |
|
Table 3 shows the comparisons and evaluations of EEEA-Net with other state-of-the-art models. EEEA-Net-C was |
|
evaluated with the test dataset, giving an error rate of 2.46% for CIFAR-10. It took 0.52 GPU days to find normal and |
|
reduce cells. |
|
By comparison, our EEEA-Net-C model achieved a lower error rate and search time than all of those other models. |
|
AmoebaNet-B was the lowest of the other state-of-the-art models: NASNet-A model, PNAS, both DARTS versions, |
|
and NSGA-Net. However, the AmoebaNet-B model required 3,150 GPU days to complete, according to Real et al. |
|
[2019]. This is clearly a hugely greater amount of search resources than required in our model (0.52 GPU days). |
|
4.1.3 Performance and Computational Complexity Analysis |
|
The multi-objective search tested error rate, number of FLOPS, and parameters. Optimisation on the effectiveness |
|
of a multi-objective uses Hypervolume (HV) as a measure of performance that computes the dominated area, using |
|
a reference point (Nadir point) with the most significant objective value from the first-generation population. Then, |
|
the Pareto-frontier solution computes the area between the reference point and Pareto. The higher HV shows that a |
|
multi-objective solution performs better in all objectives. |
|
After the model search, the HV for all the solutions obtained from the search is calculated to compare the performance |
|
of two variants: a model that uses Early Exit (EA-Net) and a model that does not use Early Exit. In Fig. 6, the values |
|
shown in the vertical axis are normalised HV , and the horizontal axis is generations. When we closely look at the HV |
|
value, it was found that the search using the EA-Net model yielded HV values greater than the three EEEA-Net models. |
|
10EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
ArchitectureCIFAR-10 Error |
|
(%)CIFAR-100 Error |
|
(%)Params |
|
(M)Search cost |
|
(GPU days)Search |
|
Method |
|
NASNet-A + CO Zoph and Le [2016] 2.83 16.58 3.1 2,000 RL |
|
ENAS + CO Pham et al. [2018] 2.89 17.27 4.6 0.5 RL |
|
PNAS Liu et al. [2018c] 3.41 17.63 3.2 225 SMBO |
|
DARTS-V1 + CO Liu et al. [2018b] 2.94 - 2.9 1.5 gradient |
|
DARTS-V2 + CO Liu et al. [2018b] 2.83 17.54 3.4 4 gradient |
|
P-DARTS + CO Chen et al. [2019] 2.50 15.92 3.4 0.3 gradient |
|
PC-DARTS + CO Xu et al. [2020] 2.57 17.36 3.6 0.1 gradient |
|
CDARTS + CO Yu and Peng [2020] 2.48 15.69 3.8 0.3 gradient |
|
AmoebaNet-A + CO Real et al. [2019] 3.12 18.93 3.1 3,150 evolution |
|
AmoebaNet-B + CO Real et al. [2019] 2.55 - 2.8 3,150 evolution |
|
LEMONADE Elsken et al. [2018] 3.05 - 4.7 80 evolution |
|
NSGA-Net + CO Lu et al. [2019] 2.75 20.74 3.3 4 evolution |
|
CARS-I + CO Yang et al. [2020] 2.62 16.00 3.6 0.4 evolution |
|
EA-Net (= 0) + CO 3.30 17.58 2.9 0.57 evolution |
|
EEEA-Net-A ( = 3) + CO 3.69 20.16 1.8 0.34 evolution |
|
EEEA-Net-B ( = 4)+ CO 2.88 16.90 1.8 0.36 evolution |
|
EEEA-Net-C ( = 5)+ CO 2.46 15.02 3.6 0.52 evolution |
|
Table 3: Comparing EEEA-Net with other architectures from RL, SMBO, gradient, and evolution search method on |
|
CIFAR-10 and CIFAR-100 datasets. |
|
0 5 10 15 20 25 30 |
|
Generation0.000.020.040.060.080.100.120.14Normalized Hypervolume |
|
EA-Net |
|
EEEA-Net-A |
|
EEEA-Net-B |
|
EEEA-Net-C |
|
Figure 6: Performance Metric of EA-Net and EEEA-Nets. |
|
However, considering only the model with an early exit, it was found that searches using equal to 5 performed |
|
better thanequal to 3 and 4, since is a parameter that determines the model size by the number of parameters. |
|
Consequently, creating larger models by increasing the size of , gave superior performance. |
|
In addition, when considering a model without an Early Exit (EA-Net) and a model that used an Early Exit ( =5, |
|
EEEA-Net-C), it was found that the search efficiency of the EEEA-Net-C model was nearly similar to that of EA-Net |
|
because the EA-Net search does not control the model size while searching for the model. Therefore, the model obtained |
|
by EA-Net’s search may is likely to obtain a model of a large parameter size. On the other hand, the model with an |
|
Early Exit better controls the model size, and the resulting model provides similar performance than achievable in an |
|
uncontrolled search. |
|
11EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
0 5 10 15 20 25 30 |
|
Generations404550556065CIFAR-10 Accuracy (%) |
|
Search space of EA-Net |
|
Pareto-front |
|
(a) CIFAR-10 Accuracy vs Generations. |
|
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 |
|
FLOPs (M)45.047.550.052.555.057.560.062.565.0CIFAR-10 Accuracy (%) |
|
Search space of EA-Net |
|
Pareto-front (b) CIFAR-10 Accuracy vs FLOPS. |
|
0 5 10 15 20 25 30 |
|
Generations404550556065CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-A |
|
Pareto-front |
|
(c) CIFAR-10 Accuracy vs Generations. |
|
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 |
|
FLOPs (M)45.047.550.052.555.057.560.062.565.0CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-A |
|
Pareto-front (d) CIFAR-10 Accuracy vs FLOPS. |
|
0 5 10 15 20 25 30 |
|
Generations404550556065CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-B |
|
Pareto-front |
|
(e) CIFAR-10 Accuracy vs Generations. |
|
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 |
|
FLOPs (M)45.047.550.052.555.057.560.062.565.0CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-B |
|
Pareto-front (f) CIFAR-10 Accuracy vs FLOPS. |
|
0 5 10 15 20 25 30 |
|
Generations404550556065CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-C |
|
Pareto-front |
|
(g) CIFAR-10 Accuracy vs Generations. |
|
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 |
|
FLOPs (M)45.047.550.052.555.057.560.062.565.0CIFAR-10 Accuracy (%) |
|
Search space of EEEA-Net-C |
|
Pareto-front (h) CIFAR-10 Accuracy vs FLOPS. |
|
Figure 7: Progress of trade-offs after each generation of EA-Net and EEEA-Nets. |
|
12EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
We present progress trade-offs after each generation of the EA-Net and EEEA-Nets search through Fig. 7. The whole |
|
population is demonstrated by two-dimensional coordinates such as CIFAR-10 accuracy vs. Generations and CIFAR-10 |
|
accuracy vs. FLOPS. |
|
4.1.4 Data augmentation |
|
The results from the evaluation of EEEA-Net are represented in precision floating-point (FP32). Our experimental goals |
|
were to create the most effective EEEA-Net without modifying the model structure. In the evaluation, we added the |
|
AutoAugment (AA) Cubuk et al. [2019] technique to the data augmentation process. AA created a more diverse set of |
|
data, making the model more effective. When Cutout DeVries and Taylor [2017] and AA Cubuk et al. [2019] were |
|
used, we observed that the error rate of EEEA-Net-C was reduced to 2.42%. Without AA, however, an error rate of |
|
2.46% occurred, as shown in Table 4. |
|
4.1.5 Architecture Evaluation on CIFAR-100 dataset |
|
The AA technique was used to optimise the search and evaluation process of the EEEA-Net. The EEEA-Net-C |
|
was trained using the CIFAR-10 dataset, which was not sufficient for our purposes. Consequently, the EEEA-Net |
|
architectures obtained from CIFAR-10 dataset were used with CIFAR-100. |
|
The hyper-parameters used in the training process were changed to evaluate the EEEA-Net with the CIFAR-100 dataset, |
|
where the number of all cells (normal and reduce cells) was set to 20 layers with 36 initial channels. This was the |
|
outcome of training the network from scratch in 600 epochs with a batch size of 128, setting the SGD optimiser with a |
|
weight decay of 0.0003 and momentum of 0.9, and the initial learning rate set to 0.025 and running with the cosine rule |
|
scheduler. The Cutout regularisation length was equal to 16, and the drop-path of probability was 0.2, with auxiliary |
|
towers of the weight of 0.4. |
|
When the EEEA-Net-C (same model structure) was evaluated with CIFAR-100 datasets, it showed an error rate of |
|
15.02%, as shown in Table 3. Further, this evaluation, with 3.6 million parameters, resulted in the lowest error rate of all |
|
the state-of-the-art models. |
|
4.2 ImageNet dataset |
|
In this subsection, we used the ImageNet dataset for the search and model evaluation. The ImageNet dataset is a |
|
large-scale standard dataset for benchmarking performance for image recognition for 1,000 classes with 1,281,167 |
|
images for the training set, 50,000 images for the test set, divided into 1,000 classes. |
|
4.2.1 Architecture Search on ImageNet |
|
Early Exit was used to discover a network architecture using the CIFAR-10 dataset. However, this network architecture |
|
was constructed from a multi-path NAS, which requires considerable memory. Given this, we used a single-path NAS |
|
to find the network architecture on ImageNet to reduce this search time, which also allows a multi-objective search |
|
with early exit population initialisation to be used on the OnceForAll Cai et al. [2020] super-network (called Supernet) |
|
to discover all network architectures that offer the best trade-off. Supernet can also search for the four dimensions |
|
of the network architecture, including kernel size, width (number of channels), depth (number of layers), and input |
|
resolution resize. We set all hyper-parameters for our architecture searches following the process in NSGA-NetV2 Lu |
|
et al. [2020a]. |
|
The two objectives of accuracy and FLOPS were the criteria for searching for 300 high accuracy samples with low |
|
FLOPS. However, these sample architectures have a diverse number of parameters. The number of parameters affects |
|
ArchitectureCIFAR-10 Error |
|
(%)Training time |
|
(GPU Hours)AutoAugment |
|
EEEA-Net-A ( = 3) + CO 3.69 25.75 - |
|
EEEA-Net-B ( = 4) + CO 2.88 25.95 - |
|
EEEA-Net-C ( = 5) + CO 2.46 48.05 - |
|
EEEA-Net-A ( = 3) + CO 3.35 30.38 Yes |
|
EEEA-Net-B ( = 4) + CO 2.87 31.26 Yes |
|
EEEA-Net-C ( = 5) + CO 2.42 54.26 Yes |
|
Table 4: Results of CIFAR-10 using Cutout (CO) and AutoAugment (AA). |
|
13EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
architecture size when running on devices that may have memory constraints. Thus, to prevent the architecture from |
|
having too many parameters, we appended the Early Exit to create the first population with limited parameters. |
|
In this experiment, we compiled the number of architecture parameters shown in Table 5 to calculate the average |
|
number of parameters equal to 5. Thus, the maximum number of parameters ( ) whereequals 5, 6 or 7, was defined |
|
as follows: EEEA-Net-A ( = 5), EEEA-Net-B ( = 6), EEEA-Net-C ( = 7). For a fair comparison, we set qual |
|
to 0, and called that EA-Net-N ( = 0). We categorised our networks using the number of MobilenetV3 FLOPS to |
|
define the network size architectures of EEEA-Net-A1, EEEA-Net-B1, EEEA-Net-C1, and EEEA-Net-N1 as small- |
|
scale architectures ( <155 FLOPS). The large-scale architectures ( <219 FLOPS) are EEEA-Net-A2, EEEA-Net-B2 |
|
EEEA-Net-C2, and EEEA-Net-N2. |
|
4.2.2 Architecture Evaluation on ImageNet dataset |
|
The discovery of architecture from Supernet is the separation of some layers from Supernet called subnets. Since |
|
Supernet and the subnets have different network architectures, the accuracy of the subnets, with pre-trained weight from |
|
Supernet, is very low when they were tested on the validation dataset. So, the subnets have calibrated the statistics of |
|
batch normalisation (BN) after searching on Supernet. The new BN statistics from the subnets were calculated using |
|
the validation dataset and updating the BN of all of the subnets. Thus, BN calibration can improve test accuracy value |
|
efficiency with the ImageNet dataset. |
|
Table 5 shows the comparison of EEEA-Net performance with other models, using three main comparison factors: error |
|
rate, number of parameters, and FLOPS. We classify models using architectural search methods such as auto, manual, |
|
or a combination. When comparing our small-scale architectures (EEEA-Net-A1, EEEA-Net-B1, EEEA-Net-C1, and |
|
EEEA-Net-N1) with GhostNet 1.0 Han et al. [2020], we found that all our architectures outperform GhostNet 1.0. |
|
Also, EEEA-Net-A1, EEEA-Net-B1, EEEA-Net-C1, and EEEA-Net-N1 provide lower error and FLOPS counts than |
|
MobileNetsV3 Large 0.75 Howard et al. [2019]. However, the MobileNetsV3 Large 0.75 has fewer parameters than our |
|
models. |
|
Similarly, when we compared our large-scale architectures with other architectures, we found that EEEA-Net-C2 |
|
(= 7) has a Top-1 error, and FLOPS were lower than all other architectures, as shown in Table 4. When we compare |
|
our architecture with MobileNetsV3 Large 1.0, EEEA-Net-C2 provides a 1% less error value than MobileNetsV3 [28], |
|
and the FLOPS count of EEEA-Net-C2 is reduced by 2 Million from MobileNetsV3. However, EEEA-Net-C2 had 0.6 |
|
Million more parameters than MobileNetsV3. |
|
Model Top-1 Error (%) Top-5 Error (%) Params (M) FLOPS (M) Type |
|
GhostNet 1.0 Han et al. [2020] 26.1 8.6 5.2 141 manual |
|
MobileNetsV3 Large 0.75 Howard et al. [2019] 26.7 - 4.0 155 combined |
|
EA-Net-N1 (ours) 26.1 8.6 4.4 140 auto |
|
EEEA-Net-A1 ( = 5)(our) 26.3 8.8 5.0 127 auto |
|
EEEA-Net-B1 ( = 6) (our) 26.0 8.5 5.0 138 auto |
|
EEEA-Net-C1 ( = 7) (our) 25.7 8.5 5.1 137 auto |
|
MobileNetsV1 Howard et al. [2017] 29.4 - 4.2 575 manual |
|
MobileNetsV2 Sandler et al. [2018] 28.0 - 3.4 300 manual |
|
GhostNet 1.3 Han et al. [2020] 24.3 7.3 7.3 226 manual |
|
MobileNetsV3 Large 1.0 Howard et al. [2019] 24.8 - 5.4 219 combined |
|
NASNet-A Zoph and Le [2016] 26.0 8.4 5.3 564 auto |
|
MnasNet-A1 Tan et al. [2019] 24.8 7.5 3.9 312 auto |
|
FBNet-C Wu et al. [2019] 25.1 - 5.5 375 auto |
|
MOGA-A Chu et al. [2020] 24.1 7.2 5.1 304 auto |
|
FairNAS-A Chu et al. [2019] 24.7 7.6 4.6 388 auto |
|
PNASNet-5 Liu et al. [2018c] 25.8 8.1 5.1 588 auto |
|
NSGANetV1-A2 Lu et al. [2020b] 25.5 8.0 4.1 466 auto |
|
OnceForAll Cai et al. [2020] 24.0 - 6.1 230 auto |
|
MSuNAS Cai et al. [2020] 24.1 - 6.1 225 auto |
|
EA-Net-N2 (ours) 24.4 7.6 5.9 226 auto |
|
EEEA-Net-A2 ( = 5) (our) 24.1 7.4 5.6 198 auto |
|
EEEA-Net-B2 ( = 6) (our) 24.0 7.5 5.7 219 auto |
|
EEEA-Net-C2 ( = 7) (our) 23.8 7.3 6.0 217 auto |
|
Table 5: Comparing EEEA-Net with other architectures from manual, combined, and auto search method on ImageNet |
|
datasets. |
|
14EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
140 160 180 200 220 |
|
FLOPs (M)73.574.074.575.075.576.0ImageNet Top-1 Accuracy (%) |
|
EA-Net |
|
EEEA-Net (=5) |
|
EEEA-Net (=6) |
|
EEEA-Net (=7) |
|
MobileNetV3 |
|
GhostNet |
|
(a) Top-1 accuracy vs FLOPS |
|
4.0 4.5 5.0 5.5 6.0 6.5 7.0 |
|
Params (M)73.574.074.575.075.576.0ImageNet Top-1 Accuracy (%) |
|
EA-Net |
|
EEEA-Net (=5) |
|
EEEA-Net (=6) |
|
EEEA-Net (=7) |
|
MobileNetV3 |
|
GhostNet (b) Top-1 accuracy vs Parameters |
|
Figure 8: Comparison of Top-1 accuracy, FLOPS (left), and parameters (right) between EEEA-Nets and MobileNetV3 |
|
[28] and GhostNet [40] on ImageNet dataset. |
|
We chose MobileNetV3 and GhostNet, including both small and large versions, to compare with our architecture, as |
|
shown in Fig. 8. Overall, we observed that EEEA-Net-C ( = 7) significantly outperforms MobileNetV3 and GhostNet |
|
for Top-1 accuracy and FLOPS. Furthermore, EEEA-Net-C ( = 7) achieves lower parameters than GhostNet. |
|
4.3 Architecture Transfer |
|
After searching and evaluating the model using the ImageNet dataset for image recognition, the models trained with the |
|
ImageNet dataset can be further developed and applied to object detection, semantic segmentation and human keypoint |
|
detection applications. |
|
4.3.1 Object detection |
|
EEEA-Net-C2 ( = 7) was used as the backbone for the object detection task to compare the effectiveness of our |
|
architecture on a real-world application. We utilised the same architecture trained with ImageNet datasets on the |
|
firmware called Single-Shot Detectors (SSD) Liu et al. [2016] and You Only Look Once version four (YOLOv4) |
|
Bochkovskiy et al. [2020]. |
|
PASCAL VOC is a standard set of data used to measure an architecture’s performance with object detection datasets. It |
|
consists of 20 classes, with bottles and plants being small objects with the lowest Average Precision (AP) of all classes. |
|
We used the SSDLite framework Sandler et al. [2018] for fast and optimised processing on mobile devices. We also |
|
used the YOLOv4 framework Bochkovskiy et al. [2020] for high precision object detection. |
|
All models were trained on the PASCAL VOC 2007 and VOC 2012 Everingham et al. [2010] train set by training the |
|
network for 200 epochs with a batch size of 32, SGD optimiser with weight decay equal to 0.0005 and momentum |
|
equal to 0.9, the initial learning rate is 0.01. It uses the scheduler with the cosine rule without a restart. All input images |
|
are resized to 320320pixels, and these models were used to evaluate the PASCAL VOC test set. |
|
For YOLOv4, we adopted MobileNet-V2 Sandler et al. [2018], MobileNet-V3 Howard et al. [2019] and EEEA-Net-C2 |
|
as the backbone of YOLOv4. All models were trained with 140 epochs and a batch size of 4. All inputs are random |
|
scales with multi-scale images ranging from 320 to 640 pixels. The label smoothing is 0.1, SGD optimiser with weight |
|
decay equal to 0.0005 and momentum equal to 0.9. The initial learning rate was set to 0.01 for a scheduler with a cosine |
|
rule with the warm-up strategy performed twice. |
|
Table 6 shows the performance of our architecture for object detection. EEEA-Net-C2 achieved a higher AP than NAS- |
|
Net, DARTS, ShuffleNet-V2, MobileNet-V2, MobileNet-V3, and MnasNet for the SSDLite framework. Nonetheless, |
|
EEEA-Net-C2 has 152 more million FLOPS than MobileNet-V3. For fairness, we used MobileNet-V2, MobileNet-V3 |
|
and EEEA-Net-C2 for training and evaluated these models using the PASCAL VOC test dataset via the YOLOv4 |
|
framework. The EEEA-Net-C2 significantly outperformed both MobileNet-V2 and MobileNet-V3. |
|
15EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Model Framework Params (M) FLOPS (M)Small Objects AP (%)VOC2007 mAP (%)Bottle Plant |
|
NASNet Zoph and Le [2016] SSDLite 5.22 1238 41.5 46.1 71.6 |
|
DARTS Liu et al. [2018b] SSDLite 4.73 1138 38.3 49.3 71.2 |
|
ShuffleNet-V2 Ma et al. [2018] SSDLite 2.17 355 29.9 38.1 65.4 |
|
MobileNet-V2 Sandler et al. [2018] SSDLite 3.30 680 37.9 43.9 69.4 |
|
MobileNet-V3 Howard et al. [2019] SSDLite 3.82 485 38.1 45.6 69.2 |
|
MnasNet Tan et al. [2019] SSDLite 4.18 708 37.7 44.4 69.6 |
|
EEEA-Net-C2 (ours) SSDLite 5.57 637 40.9 48.9 71.7 |
|
MobileNet-V2 Sandler et al. [2018] YOLOv4 46.34 8740 66.4 58.4 81.5 |
|
MobileNet-V3 Howard et al. [2019] YOLOv4 47.30 8520 68.2 50.7 78.9 |
|
EEEA-Net-C2 (ours) YOLOv4 31.15 5540 68.6 56.7 81.8 |
|
Table 6: Result of Object detection with different backbones on PASCAL VOC 2007 test set. |
|
Model Params (M) FLOPS (G) mIoU (%) |
|
NASNet Zoph and Le [2016] 7.46 36.51 77.9 |
|
DARTS Liu et al. [2018b] 6.64 34.77 77.5 |
|
ShuffleNet-V2 Ma et al. [2018] 4.10 26.30 73.0 |
|
MobileNet-V2 Sandler et al. [2018] 5.24 29.21 77.1 |
|
MobileNet-V3 Howard et al. [2019] 5.60 27.09 75.9 |
|
MnasNet Tan et al. [2019] 6.12 29.50 76.8 |
|
EEEA-Net-C2 (ours) 7.34 28.65 76.8 |
|
Table 7: Results of BiSeNet with different backbones on Cityscapes validation set. (single scale and no flipping). |
|
4.3.2 Semantic Segmentation |
|
The cityscape dataset Cordts et al. [2016] was chosen to experiment with semantic segmentation. It is a large-scale |
|
dataset of street scenes in 50 cities. Cityscapes provide dense pixel annotations of 5,000 images. These images were |
|
divided into three groups of 2,975, 500, 1,525 images for training, validation, and testing. We used BiSeNet Yu et al. |
|
[2018] with different backbones to evaluate our architecture’s performance for semantic segmentation on the Cityscapes |
|
dataset. NASNet, DARTS, ShuffleNet-V2, MobileNet-V2, MobileNet-V3, MnasNet, and EEEA-Net-C2 have trained |
|
80,000 iterations with a poly learning scheduler at an initial learning rate of 0.01 and batch size equals 16. All training |
|
images were resized to 10241024 pixels using image augmentation using colour jitter, random scale, and random |
|
horizontal flip. |
|
Table 7 shows that ShuffleNet-V2 achieved a smaller number of parameters and lower FLOPS than other architectures. |
|
However, MobileNet-V2 achieved a greater Mean Intersection over Union (mIoU) than ShuffleNet-V2, MobileNet-V3, |
|
MnasNet, and EEEA-Net-C2. The mIoU of EEEA-Net-C2 is the same as MnasNet. It is better than ShuffleNet-V2 and |
|
MobileNet-V3. |
|
4.3.3 Keypoint Detection |
|
Human keypoint detection, also known as human pose estimation, is the visual sensing of human gestures from |
|
keypoints such as the head, hips, or ankles. MS COCO Lin et al. [2014] is a comprehensive dataset to measure keypoint |
|
detection performance, consisting of data for 250,000 persons, with the data labelle at 17 keypoints. SimpleBaseline |
|
Xiao et al. [2018] is a framework for keypoint detection, enabling easier changes to backbones. Given this, it allowed |
|
us to adapt to other architectures more simply. |
|
All architectures were trained on the MS COCO train2017 set by training the network for 140 epochs with a batch size |
|
of 128, Adam optimiser, the initial learning rate is 0.001, which is reduced to 0.0001 at 90th epoch and reduced to |
|
0.00001 at 120th epoch. The training set is resized to 256192pixels using random rotation, scale, and flipping data. |
|
Table 8 shows the experimental result of SimpleBaseline with different backbones. Our EEEA-Net-C2 performed better |
|
than other backbones in the number of parameters. As well, EEEA-Net-C2 outperforms small architectures (excluding |
|
NASNet and DARTS). |
|
16EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Model Params (M) FLOPS (M) AP (%) |
|
NASNet Zoph and Le [2016] 10.66 569.11 67.9 |
|
DARTS Liu et al. [2018b] 9.20 531.77 66.9 |
|
ShuffleNet-V2 Ma et al. [2018] 7.55 154.37 60.4 |
|
MobileNet-V2 Sandler et al. [2018] 9.57 306.80 64.9 |
|
MobileNet-V3 Howard et al. [2019] 9.01 223.16 65.3 |
|
MnasNet Tan et al. [2019] 10.45 320.17 62.5 |
|
EEEA-Net-C2 (ours) 7.47 297.49 66.7 |
|
Table 8: Results of SimpleBaseline with different backbone settings on MS COCO2017 validation set. Flip is used |
|
during validation. |
|
4.4 Limitations |
|
The development of a NAS search with only one GPU processor was a challenge for the reasons set out below. Setting |
|
the appropriate number of populations, the number of generations of each population, and the number of search epochs |
|
suitable for one GPU processor presents considerable difficulties. All of these parameters affect the model search time. |
|
Increasing the number of generations increases the computing cost, but increasing the number of generations provides an |
|
opportunity for greater recombination of populations, thereby maximising the efficiency of discovering new populations. |
|
Moreover, an increased number of search epoch helps improve each population’s error fitness value. |
|
All these restrictions help to improve the NAS search. However, increasing these numbers influences search time. For |
|
example, increasing the number of search epochs from 1 epoch to 10 epochs results in a 10 increase in search time. |
|
5 Conclusion |
|
We achieved our research goals by successfully developing a CNN architecture suitable for an on-device processor with |
|
limited computing resources and applying it in real-world applications. |
|
This outcome was achieved by significantly reducing the computational cost of a neural architecture search. We |
|
introduced the Early Exit Population Initialisation (EE-PI) for Evolutionary Algorithm method to create the EEEA-Nets |
|
model. Our method achieved a massive reduction in search time on CIFAR-10 dataset; 0.34 to 0.52 GPU days. This |
|
must be seen as an outstanding outcome compared against other state-of-the-art models, such as the NSGA-Net model, |
|
which required 4 GPU days, the 2,000 GPU days of the NASNet model and the 3,150 GPU days of the AmoebaNet |
|
model. |
|
In the EEEA-Nets architecture, our emphasis was on reducing the number of parameters, the error rate and the |
|
computing cost. We were able to achieve this by introducing an Early Exit step into the Evolutionary Algorithm. |
|
Our EEEA-Nets architectures were searched on image recognition task, transferring architectures to other tasks. |
|
Experimentally, EEEA-Net-C2 is significantly better than MobileNet-V3 on image recognition, object detection, |
|
semantic segmentation, and keypoint detection tasks. Addressing this latter task had not been achieved or even |
|
attempted in any other CNN model. Therefore, our architectures can be deployed on devices with limited memory |
|
and processing capacity by achieving these significant reductions, allowing real-time processing on smartphones or |
|
on-device systems. |
|
The task of optimising the search for multi-objective evolutionary algorithms shall be continued as our future work to |
|
find better-performing models. In addition, we will consider applying a multi-objective evolutionary algorithm with |
|
EE-PI to find mobile-suitable models in other applications such as marine detection or pest detection. |
|
Acknowledgements |
|
The authors would like to acknowledge the Thailand Research Fund’s financial support through the Royal Golden |
|
Jubilee PhD. Program (Grant No. PHD/0101/2559). The study was undertaken using the National Computational |
|
Infrastructure (NCI) in Australia under the National Computational Merit Allocation Scheme (NCMAS). Further, we |
|
would like to extend our appreciation to Mr Roy I. Morien of the Naresuan University Graduate School for his assistance |
|
in editing the English grammar, syntax, and expression in the paper. |
|
17EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
References |
|
Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, |
|
Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. Imagenet large scale visual recognition |
|
challenge. International Journal of Computer Vision , 115(3):211–252, 2015. |
|
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. Imagenet classification with deep convolutional neural |
|
networks. Communications of The ACM , 60(6):84–90, 2017. |
|
Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent |
|
Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In 2015 IEEE Conference on Computer |
|
Vision and Pattern Recognition (CVPR) , pages 1–9, 2015. |
|
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In 2016 |
|
IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , pages 770–778, 2016. |
|
Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In 2018 IEEE/CVF Conference on Computer Vision |
|
and Pattern Recognition , pages 7132–7141, 2018. |
|
Forrest N. Iandola, Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, and Kurt Keutzer. Squeezenet: |
|
Alexnet-level accuracy with 50x fewer parameters and <0.5mb model size. arXiv preprint arXiv:1602.07360 , 2016. |
|
Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, |
|
and Hartwig Adam. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv |
|
preprint arXiv:1704.04861 , 2017. |
|
Chakkrit Termritthikun, Yeshi Jamtsho, and Paisarn Muneesawang. On-device facial verification using nuf-net model |
|
of deep learning. Engineering Applications of Artificial Intelligence , 85:579–589, 2019. |
|
Chakkrit Termritthikun, Yeshi Jamtsho, and Paisarn Muneesawang. An improved residual network model for image |
|
recognition using a combination of snapshot ensembles and the cutout technique. Multimedia Tools and Applications , |
|
79(1):1475–1495, 2020. |
|
Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, and Jian Sun. Shufflenet: An extremely efficient convolutional neural |
|
network for mobile devices. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages |
|
6848–6856, 2018. |
|
Barret Zoph and Quoc V . Le. Neural architecture search with reinforcement learning. In ICLR , 2016. |
|
Esteban Real, Alok Aggarwal, Yanping Huang, and Quoc V . Le. Regularized evolution for image classifier architecture |
|
search. In Proceedings of the AAAI Conference on Artificial Intelligence , volume 33, pages 4780–4789, 2019. |
|
Hanxiao Liu, Karen Simonyan, Oriol Vinyals, Chrisantha Fernando, and Koray Kavukcuoglu. Hierarchical representa- |
|
tions for efficient architecture search. In International Conference on Learning Representations , 2018a. |
|
Hanxiao Liu, Karen Simonyan, and Yiming Yang. Darts: Differentiable architecture search. In International Conference |
|
on Learning Representations , 2018b. |
|
Dimitrios Stamoulis, Ruizhou Ding, Di Wang, Dimitrios Lymberopoulos, Bodhi Priyantha, Jie Liu, and Diana |
|
Marculescu. Single-path nas: Designing hardware-efficient convnets in less than 4 hours. In Joint European |
|
Conference on Machine Learning and Knowledge Discovery in Databases , pages 481–497, 2019. |
|
Alex Krizhevsky. Learning multiple layers of features from tiny images. 2009. |
|
Mark Everingham, Luc Gool, Christopher K. Williams, John Winn, and Andrew Zisserman. The pascal visual object |
|
classes (voc) challenge. International Journal of Computer Vision , 88(2):303–338, 2010. |
|
Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, |
|
Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In 2016 IEEE |
|
Conference on Computer Vision and Pattern Recognition (CVPR) , pages 3213–3223, 2016. |
|
Tsung-Yi Lin, Michael Maire, Serge J. Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and |
|
C. Lawrence Zitnick. Microsoft coco: Common objects in context. In European Conference on Computer Vision , |
|
pages 740–755, 2014. |
|
Zhichao Lu, Ian Whalen, Vishnu Boddeti, Yashesh Dhebar, Kalyanmoy Deb, Erik Goodman, and Wolfgang Banzhaf. |
|
Nsga-net: neural architecture search using multi-objective genetic algorithm. In Proceedings of the Genetic and |
|
Evolutionary Computation Conference on , pages 419–427, 2019. |
|
Chenxi Liu, Barret Zoph, Maxim Neumann, Jonathon Shlens, Wei Hua, Li-Jia Li, Li Fei-Fei, Alan L. Yuille, Jonathan |
|
Huang, and Kevin Murphy. Progressive neural architecture search. In Proceedings of the European Conference on |
|
Computer Vision (ECCV) , pages 19–35, 2018c. |
|
18EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Hieu Pham, Melody Y . Guan, Barret Zoph, Quoc V . Le, and Jeff Dean. Efficient neural architecture search via parameter |
|
sharing. arXiv preprint arXiv:1802.03268 , 2018. |
|
Zhaohui Yang, Yunhe Wang, Xinghao Chen, Boxin Shi, Chao Xu, Chunjing Xu, Qi Tian, and Chang Xu. Cars: |
|
Continuous evolution for efficient neural architecture search. In 2020 IEEE/CVF Conference on Computer Vision |
|
and Pattern Recognition (CVPR) , pages 1829–1838, 2020. |
|
Thomas Elsken, Jan Hendrik Metzen, and Frank Hutter. Efficient multi-objective neural architecture search via |
|
lamarckian evolution. In International Conference on Learning Representations , 2018. |
|
Mingxing Tan and Quoc V . Le. Efficientnet: Rethinking model scaling for convolutional neural networks. In |
|
International Conference on Machine Learning , pages 6105–6114, 2019. |
|
Alvin Wan, Xiaoliang Dai, Peizhao Zhang, Zijian He, Yuandong Tian, Saining Xie, Bichen Wu, Matthew Yu, Tao Xu, |
|
Kan Chen, Peter Vajda, and Joseph E. Gonzalez. Fbnetv2: Differentiable neural architecture search for spatial and |
|
channel dimensions. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages |
|
12965–12974, 2020. |
|
Xin Chen, Lingxi Xie, Jun Wu, and Qi Tian. Progressive differentiable architecture search: Bridging the depth gap |
|
between search and evaluation. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV) , pages |
|
1294–1303, 2019. |
|
Hongyuan Yu and Houwen Peng. Cyclic differentiable architecture search. arXiv preprint arXiv:2006.10724 , 2020. |
|
Zhichao Lu, Kalyanmoy Deb, Erik D. Goodman, Wolfgang Banzhaf, and Vishnu Naresh Boddeti. Nsganetv2: |
|
Evolutionary multi-objective surrogate-assisted neural architecture search. In European Conference on Computer |
|
Vision , pages 35–51, 2020a. |
|
Andrew Howard, Ruoming Pang, Hartwig Adam, Quoc Le, Mark Sandler, Bo Chen, Weijun Wang, Liang-Chieh Chen, |
|
Mingxing Tan, Grace Chu, Vijay Vasudevan, and Yukun Zhu. Searching for mobilenetv3. In 2019 IEEE/CVF |
|
International Conference on Computer Vision (ICCV) , pages 1314–1324, 2019. |
|
Yuhui Xu, Lingxi Xie, Xiaopeng Zhang, Xin Chen, Guo-Jun Qi, Qi Tian, and Hongkai Xiong. Pc-darts: Partial channel |
|
connections for memory-efficient architecture search. In ICLR 2020 : Eighth International Conference on Learning |
|
Representations , 2020. |
|
Mingxing Tan, Bo Chen, Ruoming Pang, Vijay Vasudevan, Mark Sandler, Andrew Howard, and Quoc V . Le. Mnasnet: |
|
Platform-aware neural architecture search for mobile. In 2019 IEEE/CVF Conference on Computer Vision and |
|
Pattern Recognition (CVPR) , pages 2820–2828, 2019. |
|
Lingxi Xie and Alan Yuille. Genetic cnn. In 2017 IEEE International Conference on Computer Vision (ICCV) , 2017. |
|
Alejandro Baldominos, Yago Saez, and Pedro Isasi. Evolutionary convolutional neural networks: An application to |
|
handwriting recognition. Neurocomputing , 283:38–52, 2017. |
|
Esteban Real, Sherry Moore, Andrew Selle, Saurabh Saxena, Yutaka Leon Suematsu, Jie Tan, Quoc V . Le, and Alexey |
|
Kurakin. Large-scale evolution of image classifiers. In ICML’17 Proceedings of the 34th International Conference |
|
on Machine Learning - Volume 70 , pages 2902–2911, 2017. |
|
K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan. A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE |
|
Transactions on Evolutionary Computation , 6(2):182–197, 2002. |
|
Jesús Velasco Carrau, Gilberto Reynoso-Meza, Sergio García-Nieto, and Xavier Blasco. Enhancing controller’s |
|
tuning reliability with multi-objective optimisation: From model in the loop to hardware in the loop. Engineering |
|
Applications of Artificial Intelligence , 64:52–66, 2017. |
|
Mahmudul Hasan, Khin Lwin, Maryam Imani, Antesar M. Shabut, Luiz F. Bittencourt, and Mohammed Alamgir |
|
Hossain. Dynamic multi-objective optimisation using deep reinforcement learning: benchmark, algorithm and an |
|
application to identify vulnerable zones based on water quality. Engineering Applications of Artificial Intelligence , |
|
86:107–135, 2019. |
|
Ekin D. Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V . Le. Autoaugment: Learning augmentation |
|
strategies from data. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages |
|
113–123, 2019. |
|
Terrance DeVries and Graham W. Taylor. Improved regularization of convolutional neural networks with cutout. arXiv |
|
preprint arXiv:1708.04552 , 2017. |
|
Kai Han, Yunhe Wang, Qi Tian, Jianyuan Guo, Chunjing Xu, and Chang Xu. Ghostnet: More features from cheap |
|
operations. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages 1580–1589, |
|
2020. |
|
19EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted |
|
residuals and linear bottlenecks. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages |
|
4510–4520, 2018. |
|
Bichen Wu, Kurt Keutzer, Xiaoliang Dai, Peizhao Zhang, Yanghan Wang, Fei Sun, Yiming Wu, Yuandong Tian, Peter |
|
Vajda, and Yangqing Jia. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. |
|
In2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages 10734–10742, 2019. |
|
Xiangxiang Chu, Bo Zhang, and Ruijun Xu. Moga: Searching beyond mobilenetv3. In ICASSP 2020 - 2020 IEEE |
|
International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pages 4042–4046, 2020. |
|
Xiangxiang Chu, Bo Zhang, Ruijun Xu, and Jixiang Li. Fairnas: Rethinking evaluation fairness of weight sharing |
|
neural architecture search. arXiv preprint arXiv:1907.01845 , 2019. |
|
Zhichao Lu, Ian Whalen, Yashesh Dhebar, Kalyanmoy Deb, Erik Goodman, Wolfgang Banzhaf, and Vishnu Naresh |
|
Boddeti. Multi-objective evolutionary design of deep convolutional neural networks for image classification. IEEE |
|
Transactions on Evolutionary Computation , pages 1–1, 2020b. |
|
Han Cai, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. Once for all: Train one network and specialize it |
|
for efficient deployment. In ICLR 2020 : Eighth International Conference on Learning Representations , 2020. |
|
Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott E. Reed, Cheng-Yang Fu, and Alexander C. |
|
Berg. Ssd: Single shot multibox detector. In 14th European Conference on Computer Vision, ECCV 2016 , pages |
|
21–37, 2016. |
|
Alexey Bochkovskiy, Chien-Yao Wang, and Hong-Yuan Mark Liao. Yolov4: Optimal speed and accuracy of object |
|
detection. arXiv preprint arXiv:2004.10934 , 2020. |
|
Ningning Ma, Xiangyu Zhang, Hai-Tao Zheng, and Jian Sun. Shufflenet v2: Practical guidelines for efficient cnn |
|
architecture design. In Proceedings of the European Conference on Computer Vision (ECCV) , pages 122–138, 2018. |
|
Changqian Yu, Jingbo Wang, Chao Peng, Changxin Gao, Gang Yu, and Nong Sang. Bisenet: Bilateral segmentation |
|
network for real-time semantic segmentation. In Proceedings of the European Conference on Computer Vision |
|
(ECCV) , pages 325–341, 2018. |
|
Bin Xiao, Haiping Wu, and Yichen Wei. Simple baselines for human pose estimation and tracking. In Proceedings of |
|
the European Conference on Computer Vision (ECCV) , pages 472–487, 2018. |
|
Chris Ying, Aaron Klein, Eric Christiansen, Esteban Real, Kevin Murphy, and Frank Hutter. Nas-bench-101: Towards |
|
reproducible neural architecture search. In International Conference on Machine Learning , pages 7105–7114, 2019. |
|
Arber Zela, Julien Siems, and Frank Hutter. Nas-bench-1shot1: Benchmarking and dissecting one-shot neural |
|
architecture search. In ICLR 2020 : Eighth International Conference on Learning Representations , 2020. |
|
Xuanyi Dong and Yi Yang. Nas-bench-201: Extending the scope of reproducible neural architecture search. In Eighth |
|
International Conference on Learning Representations , 2020. |
|
Liam Li and Ameet Talwalkar. Random search and reproducibility for neural architecture search. In Uncertainty in |
|
Artificial Intelligence , pages 367–377, 2019. |
|
Ronald J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine |
|
Learning , 8(3):229–256, 1992. |
|
Xuanyi Dong and Yi Yang. Searching for a robust neural architecture in four gpu hours. In 2019 IEEE/CVF Conference |
|
on Computer Vision and Pattern Recognition (CVPR) , pages 1761–1770, 2019. |
|
Sirui Xie, Hehui Zheng, Chunxiao Liu, and Liang Lin. Snas: stochastic neural architecture search. In International |
|
Conference on Learning Representations , 2018. |
|
Shoukang Hu, Sirui Xie, Hehui Zheng, Chunxiao Liu, Jianping Shi, Xunying Liu, and Dahua Lin. Dsnas: Direct neural |
|
architecture search without parameter retraining. In 2020 IEEE/CVF Conference on Computer Vision and Pattern |
|
Recognition (CVPR) , pages 12084–12092, 2020. |
|
6 Appendix |
|
6.1 Architecture Visualisation |
|
This section visualises the architectures obtained by searching for EEEA-Nets with CIFAR-10 datasets, as shown in |
|
Fig. 9, and EEEA-Nets with ImageNet datasets shown in Fig. 10. These architectures are the most reliable, minimising |
|
three goals. |
|
20EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
c_{k-2} 0dil_conv_3x3 |
|
max_pool_3x3 |
|
c_{k-1}1inv_res_5x5 |
|
dil_conv_5x5 |
|
2avg_pool_3x3 |
|
avg_pool_3x33inv_res_3x3 |
|
c_{k}inv_res_5x5 |
|
4max_pool_3x3 |
|
sep_conv_3x3 |
|
(a) EA-Net ( = 0) Normal Cell. |
|
c_{k-2}0avg_pool_3x3 |
|
inv_res_3x3 |
|
1dil_conv_3x32inv_res_3x3 |
|
dil_conv_5x54max_pool_3x3sep_conv_3x3 |
|
c_{k-1}inv_res_5x5c_{k} |
|
3max_pool_3x3 |
|
skip_connect (b) EA-Net ( = 0) Reduction Cell. |
|
c_{k-2}0max_pool_3x3 |
|
inv_res_3x3 |
|
1max_pool_3x32inv_res_3x3 |
|
c_{k-1} dil_conv_3x3skip_connect |
|
3avg_pool_3x3 |
|
c_{k} |
|
skip_connect |
|
4skip_connectavg_pool_3x3 |
|
(c) EEEA-Net-A ( =3.0) Normal Cell. |
|
c_{k-2}0 |
|
skip_connectdil_conv_3x3 |
|
1avg_pool_3x32 dil_conv_3x3avg_pool_3x3 |
|
4dil_conv_5x5 |
|
c_{k-1}max_pool_3x33skip_connect |
|
c_{k}skip_connect |
|
dil_conv_3x3 (d) EEEA-Net-A ( =3.0) Reduction Cell. |
|
c_{k-2} 0skip_connect |
|
max_pool_3x3 |
|
c_{k-1} |
|
1max_pool_3x3skip_connect2 skip_connect |
|
3sep_conv_3x3skip_connect |
|
c_{k} |
|
dil_conv_3x34max_pool_3x3 |
|
sep_conv_3x3 |
|
(e) EEEA-Net-B ( =4.0) Normal Cell. |
|
c_{k-2} 0skip_connectinv_res_5x5 |
|
1avg_pool_3x3 |
|
c_{k-1}dil_conv_3x3 |
|
2 |
|
inv_res_3x3skip_connect3max_pool_3x3 |
|
avg_pool_3x34avg_pool_3x3 |
|
c_{k}dil_conv_3x3 (f) EEEA-Net-B ( =4.0) Reduction Cell. |
|
c_{k-2}0dil_conv_5x5 |
|
max_pool_3x3 |
|
1inv_res_5x5 |
|
inv_res_3x3c_{k-1} |
|
3sep_conv_5x54sep_conv_3x32 |
|
dil_conv_5x5inv_res_3x3 |
|
skip_connectc_{k}avg_pool_3x3 |
|
(g) EEEA-Net-C ( =5.0) Normal Cell. |
|
c_{k-2} |
|
0inv_res_3x3 |
|
dil_conv_3x31avg_pool_3x3 |
|
inv_res_3x3 |
|
c_{k-1} 2dil_conv_5x5 |
|
max_pool_3x34inv_res_3x3c_{k}3skip_connect |
|
max_pool_3x3 |
|
inv_res_5x5 (h) EEEA-Net-C ( =5.0) Reduction Cell. |
|
Figure 9: Normal and Reduction cells learned on CIFAR-10: EA-Net ( =0.0), EEEA-Net-A ( =3.0), EEEA-Net-B |
|
(=4.0), and EEEA-Net-C ( =5.0). |
|
21EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
K=3 |
|
E=3 K=3 |
|
E=4 K=3 |
|
E=6 K=5 |
|
E=3 K=5 |
|
E=4 K=5 |
|
E=6 K=7 |
|
E=3 K=7 |
|
E=4 K=7 |
|
E=6 Skip Legend |
|
Stem Tail Stem Tail |
|
Stem TailStage 1 Stage 2 Stage 3 Stage 4 Stage 5 |
|
EA-Net-N1 EA-Net-N2 |
|
Stem TailEEEA-Net-A1 |
|
( 𝛽 = 5) EEEA-Net-A2 |
|
( 𝛽 = 5) |
|
Stem Tail Stem Tail |
|
Stem TailEEEA-Net-B1 |
|
( 𝛽 = 6) EEEA-Net-B2 |
|
( 𝛽 = 6) |
|
Stem TailEEEA-Net-C1 |
|
( 𝛽 = 7) EEEA-Net-C2 |
|
( 𝛽 = 7) Stage 1 Stage 2 Stage 3 Stage 4 Stage 5 |
|
Figure 10: EA-Nets and EEEA-Nets architectures were searched from ImageNet datasets. The stem and tail layers in |
|
all architectures are the same. |
|
EEEA -Net-C2 |
|
MobileNetv 3 |
|
MobileNetv 2 |
|
Figure 11: An example of object detection results of MobileNetv2, MobileNetv3, and EEEA-Net-C2 models. |
|
6.2 Error Analysis of EEEA-Net-C2 |
|
Our experiment applied EEEA-Net-C2 for detection, semantic segmentation, and human keypoint detection, where |
|
we concluded that the EEEA-Net-C2 model was better than the MobileNet-V3 model. For error analysis of the |
|
EEEA-Net-C2 model, images for each application were processed to check the correct results. In this appendix, error |
|
analysis of the EEEA-Net-C2 model is divided into three parts: error analysis of object detection, error analysis of |
|
semantic segmentation errors and error analysis of human keypoint detection. |
|
6.2.1 Object detection |
|
Object detection results from MobileNetv2, MobileNetv3 and EEEA-Net-C2 models are shown in Fig. 11. Given the |
|
error from the images in the first column, the MobileNetv2 model was found to have mistakenly identified “bird" as |
|
“person", while the MobileNetv3 model was unable to find “bird". However, the EEEA-Net-C2 model detected the |
|
location of all “birds". |
|
As with the second column in Fig. 11, the EEEA-Net-C2 model can identify all “persons" positions. However, only |
|
the EEEA-Net-C2 model could not locate the hidden “person" behind the middle woman in the third column image. |
|
Additionally, in the fourth column pictures, the EEEA-Net-C2 model was able to identify more “plant pots" than the |
|
MobileNetv2 and MobileNetv3 models. |
|
22EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
MobileNetv 3 |
|
EEEA -Net-C2 |
|
Input |
|
Figure 12: An example of semantic segmentation results of MobileNetv2, MobileNetv3 and EEEA-Net-C2 models. |
|
MobileNetv 3 |
|
EEEA -Net-C2MobileNetv 2 |
|
Figure 13: An example of human keypoint detection results of MobileNetv2, MobileNetv3 and EEEA-Net-C2 models. |
|
6.2.2 Semantic segmentation |
|
The results of visual image segmentation using MobileNetv3 and EEEA-Net-C2 models are shown in Fig. 12. The |
|
error of the semantic segmentation results can be determined from the pictures of the first column. It was found that |
|
MobileNetv3 models could segment only “Traffic sign pole". However, the MobileNetv3 model cannot segment the left |
|
“traffic sign", while the EEEA-Net-C2 model can segment both “pole and sign". |
|
The pictures in the second column from Fig. 12 depicts that the EEEA-Net-C2 model segmented from the “traffic |
|
island" was less than the MobileNetv3 model. Next, in the third column, the EEEA-Net-C2 model segmented the |
|
“footpath" more precisely than the MobileNetv3 model. |
|
6.2.3 Human keypoint detection |
|
The human keypoint detection results from the MobileNetv2, MobileNetv3 and EEEA-Net-C2 models are shown in |
|
Fig. 13. When considering the error from the pictures in the first column, the MobileNetv3 model was found to indicate |
|
the “left arm" position, while the MobileNetv2 and EEEA-Net-C2 models were able to locate. |
|
23EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
Figure 14: Comparison of latency between EEEA-Net and other state-of-the-art models on non-GPU processing. |
|
Only the MobileNetv3 model can pinpoint the “leg" of the person sitting in the second column pictures. However, in the |
|
third column pictures, the EEEA-Net-C2 model can locate the middle person’s “arms and legs", while the MobileNetv3 |
|
model identifies the person’s wrong location. Additionally, in the fourth column pictures, the EEEA-Net-C2 model |
|
could locate the “arm and leg" more accurately than the MobileNetv2 and MobileNetv3 models. |
|
The above data shows that the EEEA-Net-C2 model is accurate as well as inaccurate. The EEEA-Net-C2 model was |
|
designed and searched with the ImageNet dataset. Thus, the EEEA-Net-C2 model may have errors when used with other |
|
dataset or tasks. However, the EEEA-Net-C2 model has a performance higher than the MobileNetv2 and MobileNetv3 |
|
models on the same dataset and framework used in the three applications. |
|
6.3 Mobile Processing |
|
This appendix measures the performance of our EEEA-Net-C2 model and other state-of-the-art models on the smart- |
|
phone and only CPU. All trained models with the ImageNet dataset are converted to the PyTorch JIT version to enable |
|
easy implementation on different platforms. |
|
Fig. 14 shows the latency performance with 100 images with 224x224 pixels on the Google Pixel 3 XL smartphone |
|
(blue bars) and Intel i7-6700HQ CPU (red bars) devices with non-GPU resources by DARTSv2, P-DARTS, NASNet, |
|
ReletiveNAS, ShuffleNetV2, MNASNet 1.0, MobileNetV2, MobileNetV3, and EEEA-Net-C2 models. |
|
On the Google Pixel 3 XL, the EEEA-Net-C2 model processed each image in 86 milliseconds per image, whereas the |
|
MobileNetV2 and MobileNetV3 models took 90 and 88 milliseconds, respectively. The EEEA-Net-C2 model has a |
|
shorter latency time than state-of-the-art models (including DARTSv2, P-DARTS, NASNet, and ReletiveNAS), and |
|
MobileNets models, primarily models for smartphones, are compared to EEEA-Net-C2. |
|
Also, on the Intel i7-6700HQ CPU, the latency time of the EEEA-Net-C2 model has shorter latency than state-of-the-art |
|
models and lightweight models (including MNASNet 1.0, MobileNetV2, and MobileNetV3). |
|
6.4 NAS-Bench dataset |
|
Experimental results with CIFAR-10, CIFAR-100, and ImageNet datasets were compared between NAS methods. The |
|
results obtained from different methods are achieved with different settings such as hyperparameters (e.g., learning rate |
|
and batch size), data augmentation (e.g., Cutout and AutoAugment). Thus, the comparison may not be fair. |
|
24EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
0 1 2 3 4 5 |
|
total training time spent (seconds) 1e60.9360.9380.9400.9420.944accuracy |
|
random |
|
evolution |
|
early exit evolution |
|
Figure 15: Comparison of accuracy between random search, regularised evolution and Early Exit evolution algorithms |
|
on NAS-Bench-101. |
|
This section has implemented an Early Exit method to model search from the NAS-Bench datasets, which avoids |
|
unfair comparisons and provides a uniform benchmark for NAS algorithms. The dataset used in this experiment was |
|
NAS-Bench-101, NAS-Bench-1Shot1 and NAS-Bench-201. |
|
6.4.1 NAS-Bench-101 |
|
The NAS-Bench-101 Ying et al. [2019] provides a table dataset of 423,624 unique architectures. These architectures |
|
have trained and evaluated the CIFAR-10 dataset to allow our work to search and query the mapped performance in the |
|
dataset in a few milliseconds. |
|
We have re-implemented model search from the NAS-Bench-101 dataset by using random search, regularised evolution |
|
and Early Exit evolution algorithms to search and query the performance of the resulting models. We used a re- |
|
implemented regularised evolution with the Early Exit method by taking population size of 100, a tournament size of |
|
10, and maximum parameters’ Early Exit ( ) is 25 million. |
|
The results in Fig. 15 show that our early exit evolution algorithm tends to be higher in accuracy than the regularised |
|
evolution from 2 million seconds to 5 million seconds. Overall, the regularised evolution algorithm appears to perform |
|
better than the random search algorithm. However, our early exit evolution tends to outperform both random search and |
|
regularised evolution algorithms. |
|
6.4.2 NAS-Bench-1Shot1 |
|
NAS-Bench-1Shot1 Zela et al. [2020] is the benchmark for a one-shot neural architecture search, developed from the |
|
NAS-Bench-101 search space by tracking the trajectory and performance of the obtained architectures for three search |
|
spaces: 6,240 architectures for search space 1, 29,160 architectures for search space 2, and 363,648 architectures for |
|
search space 3. |
|
Fig. 16 shows the mean performance on validation regret of architectures obtained by random search, regularised |
|
evolution and Early Exit evolution algorithms. For search space 1, our algorithm achieves validation regret close to |
|
the regularised evolution algorithm. For search space 2, our algorithm converges better than the regularised evolution |
|
algorithm. Our algorithm outperforms random search and regularised evolution algorithms for search space 3, the most |
|
significant (100more architectures than space 1, and 10 more than space 2). |
|
25EEEA: Early Exit Evolutionary Algorithm Network by Chakkrit Termritthikun, et al. |
|
104 |
|
105 |
|
estimated wallclock time [s]102 |
|
validation regret |
|
search space 1 |
|
RS |
|
RE |
|
EE |
|
(a) |
|
104 |
|
105 |
|
estimated wallclock time [s]102 |
|
validation regret |
|
search space 2 |
|
RS |
|
RE |
|
EE (b) |
|
104 |
|
105 |
|
estimated wallclock time [s]102 |
|
6×103 |
|
2×102 |
|
validation regret |
|
search space 3 |
|
RS |
|
RE |
|
EE (c) |
|
Figure 16: Comparison of accuracy between random search (RS), regularised evolution (RE) and Early Exit evolution |
|
(EE) algorithms on NAS-Bench-1Shot1. |
|
MethodCIFAR-10 CIFAR-100 ImageNet16-120 |
|
validation test validation test validation test |
|
ResNet He et al. [2016] 90.83 93.97 70.42 70.86 44.53 43.63 |
|
RSPS Li and Talwalkar [2019] 84.16±1.69 87.66±1.69 45.78±6.33 46.60±6.57 31.09±5.65 30.78±6.12 |
|
Reinforce Williams [1992] 91.09±0.37 93.85±0.37 70.05±1.67 70.17±1.61 43.04±2.18 43.16±2.28 |
|
ENAS Pham et al. [2018] 39.77±0.00 54.30±0.00 10.23±0.12 10.62±0.27 16.43±0.00 16.32±0.00 |
|
DARTS Liu et al. [2018b] 39.77±0.00 54.30±0.00 38.57±0.00 38.97±0.00 18.87±0.00 18.41±0.00 |
|
GDAS Dong and Yang [2019] 90.01±0.46 93.23±0.23 24.05±8.12 24.20±8.08 40.66±0.00 41.02±0.00 |
|
SNAS Xie et al. [2018] 90.10±1.04 92.77±0.83 69.69±2.39 69.34±1.98 42.84±1.79 43.16±2.64 |
|
DSNAS Hu et al. [2020] 89.66±0.29 93.08±0.13 30.87±16.40 31.01±16.38 40.61±0.09 41.07±0.09 |
|
PC-DARTS Xu et al. [2020] 89.96±0.15 93.41±0.30 67.12±0.39 67.48±0.89 40.83±0.08 41.31±0.22 |
|
EA-Net (SO) 91.53±0.00 94.22±0.00 73.13±0.00 73.17±0.00 46.32±0.00 46.48±0.00 |
|
EA-Net (= 0) 88.97±2.48 91.54±2.69 66.84±5.08 67.00±4.90 39.93±5.54 39.27±6.21 |
|
EEEA-Net ( = 0.3) 87.07±1.59 89.76±1.87 64.04±3.21 64.31±3.21 35.42±3.81 34.98±4.13 |
|
EEEA-Net ( = 0.4) 89.91±0.77 92.68±0.69 68.70±1.50 68.65±1.51 41.71±1.58 41.25±1.61 |
|
EEEA-Net ( =0.5) 90.21±0.58 92.83±0.46 69.15±1.36 68.95±1.25 42.14±1.14 41.98±1.22 |
|
Optimal 91.61 94.37 73.49 73.51 46.77 47.31 |
|
Table 9: Comparison of a single objective and multi-objective evolution algorithm with the 8 NAS methods provided by |
|
NAS-Bench-201 benchmark. Optimal shows the best architecture in the search space. |
|
6.4.3 NAS-Bench-201 |
|
NAS-Bench-201 Dong and Yang [2020] is an extension of NAS-Bench-101, which extends different search spaces, |
|
and it has a wide range of datasets, including CIFAR-10, CIFAR-100, and ImageNet-16-120. It contains 15,625 |
|
architectures by five operations, and 6-dimensional vectors indicate the operation in the cell. All architecture evaluated |
|
performance by validation and test sets on CIFAR-10, CIFAR-100, and ImageNet-16-120. |
|
We compare our Early Exit Evolution Algorithm or EEEA-Net ( = 0.3, 0.4 and 0.5) with the single objective evolution |
|
algorithms (SO) and multi-objective evolution algorithms ( = 0). The hyper-parameters for this search process were |
|
defined as the generations of EA equal to 10 generations with 100 populations by retaining a probability of 0.5, a |
|
mutation probability of 0.1, and Early Exit is the maximum number of parameters equal to 0.3, 0.4 and 0.5 million. |
|
The results are shown in Table 9, our EEEA-Net ( = 0.4 and 0.5) outperforms EEEA-Net ( = 0 and 0.3). However, |
|
the EA-Net (SO) using accuracy as the optimisation objective performed better than all EEEA-Nets. |
|
Furthermore, when we compared our EEEA-Net ( = 0.5) with 8 NAS methods, including RSPS Li and Talwalkar |
|
[2019], Reinforce Williams [1992], ENAS Pham et al. [2018], DARTS Liu et al. [2018b], GDAS Dong and Yang |
|
[2019], SNAS Xie et al. [2018], DSNAS, and PC-DARTS Xu et al. [2020], we found that EEEA-Net ( = 0.5) has an |
|
accuracy was higher than all other NAS method except for the Reinforce method. |
|
26 |