A Novel Training Protocol for Performance Predictors of Evolutionary Neural Architecture Search Algorithms | IEEE Journals & Magazine | IEEE Xplore

A Novel Training Protocol for Performance Predictors of Evolutionary Neural Architecture Search Algorithms


Abstract:

Evolutionary neural architecture search (ENAS) can automatically design the architectures of deep neural networks (DNNs) using evolutionary computation algorithms. Howeve...Show More

Abstract:

Evolutionary neural architecture search (ENAS) can automatically design the architectures of deep neural networks (DNNs) using evolutionary computation algorithms. However, most ENAS algorithms require an intensive computational resource, which is not necessarily available to the users interested. Performance predictors are a type of regression models which can assist to accomplish the search, while without exerting much computational resource. Despite various performance predictors have been designed, they employ the same training protocol to build the regression models: 1) sampling a set of DNNs with performance as the training dataset; 2) training the model with the mean square error criterion; and 3) predicting the performance of DNNs newly generated during the ENAS. In this article, we point out that the three steps constituting the training protocol are not well thought-out through intuitive and illustrative examples. Furthermore, we propose a new training protocol to address these issues, consisting of designing a pairwise ranking indicator to construct the training target, proposing to use the logistic regression to fit the training samples, and developing a differential method to build the training instances. To verify the effectiveness of the proposed training protocol, four widely used regression models in the field of machine learning have been chosen to perform the comparisons on two benchmark datasets. The experimental results of all the comparisons demonstrate that the proposed training protocol can significantly improve the performance prediction accuracy against the traditional training protocols.
Published in: IEEE Transactions on Evolutionary Computation ( Volume: 25, Issue: 3, June 2021)
Page(s): 524 - 536
Date of Publication: 27 January 2021

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Deep neural networks (DNNs) are becoming the dominant algorithm of machine learning [1], largely owing to their superiority in solving challenging real-world applications [2], [3]. Generally, the performance of DNNs relies on two deciding factors: 1) the architectures of the DNNs and 2) the weights associated with the architecture. The performance of a DNN in solving the corresponding problem can be promising, only when its architecture and the weights achieve the optimum combination simultaneously. Commonly, when the architecture of a DNN is determined, the optimal weights can be obtained through formulizing the loss as a continuous function, and then the exact optimization algorithms are employed for solving. In practice, the gradient-based optimization algorithms are the most popular ones in addressing the loss function, although they cannot theoretically guarantee the global optimum [4]. On the other hand, obtaining the optimal architectures is not a trivial task because the architectures cannot be directly optimized as the weights do. In practice, most, if not all, prevalent state-of-the-art DNN architectures are manually designed based on extensive human expertise, including ResNet [5], DenseNet [6], and among others.

Select All
1.
Y. LeCun, Y. Bengio and G. Hinton, "Deep learning", Nature, vol. 521, no. 7553, pp. 436-444, 2015.
2.
G. E. Hinton and R. R. Salakhutdinov, "Reducing the dimensionality of data with neural networks", Science, vol. 313, no. 5786, pp. 504-507, 2006.
3.
A. Krizhevsky, I. Sutskever and G. E. Hinton, "ImageNet classification with deep convolutional neural networks" in Advances in Neural Information Processing Systems, Red Hook, NY, USA:Curran, pp. 1097-1105, 2012.
4.
J. K. Kearney, W. B. Thompson and D. L. Boley, "Optical flow estimation: An error analysis of gradient-based methods with local optimization", IEEE Trans. Pattern Anal. Mach. Intell., vol. PAMI-9, no. 2, pp. 229-244, Mar. 1987.
5.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 770-778, 2016.
6.
G. Huang, Z. Liu, K. Q. Weinberger and L. van der Maaten, "Densely connected convolutional networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2261-2269, 2017.
7.
T. Elsken, J. H. Metzen and F. Hutter, Neural architecture search: A survey, 2018, [online] Available: http://arXiv:1808.05377.
8.
Y. Sun, B. Xue, M. Zhang and G. G. Yen, "Completely automated CNN architecture design based on blocks", IEEE Trans. Neural Netw. Learn. Syst., vol. 31, no. 4, pp. 1242-1254, Apr. 2020.
9.
B. Baker, O. Gupta, N. Naik and R. Raskar, "Designing neural network architectures using reinforcement learning", Proc. Int. Conf. Learn. Represent., 2017, [online] Available: https://openreview.net/pdf?id=S1c2cvqee.
10.
B. Zoph and Q. V. Le, "Neural architecture search with reinforcement learning", Proc. Int. Conf. Learn. Represent., pp. 1-16, 2017.
11.
H. Liu, K. Simonyan and Y. Yang, DARTS: Differentiable architecture search, 2018, [online] Available: http://arXiv:1806.09055.
12.
L. Xie and A. Yuille, "Genetic CNN", Proc. IEEE Int. Conf. Comput. Vis., pp. 1388-1397, 2017.
13.
E. Real et al., "Large-scale evolution of image classifiers", Proc. 34th Int. Conf. Mach. Learn. Res., pp. 2902-2911, 2017.
14.
H. Liu, K. Simonyan, O. Vinyals, C. Fernando and K. Kavukcuoglu, "Hierarchical representations for efficient architecture search", Proc. Int. Conf. Mach. Learn. Res., 2018, [online] Available: https://openreview.net/pdf?id=BJQRKzbA-.
15.
M. Suganuma, S. Shirakawa and T. Nagao, "A genetic programming approach to designing convolutional neural network architectures", Proc. Genet. Evol. Comput. Conf., pp. 497-504, 2017.
16.
Y. Sun, B. Xue, M. Zhang and G. G. Yen, "Evolving deep convolutional neural networks for image classification", IEEE Trans. Evol. Comput., vol. 24, no. 2, pp. 394-407, Apr. 2020.
17.
Z. Li, T. Xi, J. Deng, G. Zhang, S. Wen and R. He, "GP-NAS: Gaussian process based neural architecture search", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 11933-11942, 2020.
18.
L. Ma, J. Cui and B. Yang, "Deep neural architecture search with deep graph Bayesian optimization", Proc. IEEE/WIC/ACM Int. Conf. Web Intell. (WI), pp. 500-507, 2019.
19.
K. Kandasamy, W. Neiswanger, J. Schneider, B. Poczos and E. P. Xing, "Neural architecture search with Bayesian optimisation and optimal transport" in Advances in Neural Information Processing Systems, Red Hook, NY, USA:Curran, pp. 2016-2025, 2018.
20.
T. D. Ottelander, A. Dushatskiy, M. Virgolin and P. A. N. Bosman, Local search is a remarkably strong baseline for neural architecture search, 2020, [online] Available: http://arXiv:2004.08996.
21.
L. Li and A. Talwalkar, "Random search and reproducibility for neural architecture search", Proc. 35th Uncertainty Artif. Intell. Conf., pp. 367-377, 2020.
22.
T. Back, Evolutionary Algorithms in Theory and Practice: Evolution Strategies Evolutionary Programming Genetic Algorithms, London, U.K.:Oxford Univ. Press, 1996.
23.
W. Banzhaf, P. Nordin, R. E. Keller and F. D. Francone, Genetic Programming: An Introduction, San Francisco, CA, USA:Morgan Kaufmann, 1998.
24.
L. M. Schmitt, "Theory of genetic algorithms", Theor. Comput. Sci., vol. 259, no. 1, pp. 1-61, 2001.
25.
H. Wang, L. Jiao and X. Yao, "Two_arch2: An improved two-archive algorithm for many-objective optimization", IEEE Trans. Evol. Comput., vol. 19, no. 4, pp. 524-541, Aug. 2015.
26.
Y. Sun, G. G. Yen and Z. Yi, "IGD indicator-based evolutionary algorithm for many-objective optimization problems", IEEE Trans. Evol. Comput., vol. 23, no. 2, pp. 173-187, Apr. 2019.
27.
Y. Sun, B. Xue, M. Zhang and G. G. Yen, "A new two-stage evolutionary algorithm for many-objective optimization", IEEE Trans. Evol. Comput., vol. 23, no. 5, pp. 748-761, Oct. 2019.
28.
B. Wang, Y. Sun, B. Xue and M. Zhang, "Evolving deep convolutional neural networks by variable-length particle swarm optimization for image classification", Proc. IEEE Congr. Evol. Comput. (CEC), pp. 1-8, 2018.
29.
N. Qiang, B. Ge, Q. Dong, F. Ge and T. Liu, "Neural architecture search for optimizing deep belief network models of FMRI data", Proc. Int. Workshop Multiscale Multimodal Med. Imag., pp. 26-34, 2019.
30.
J. Hajewski and S. Oliveira, "An evolutionary approach to variational autoencoders", Proc. IEEE 10th Annu. Comput. Commun. Workshop Conf. (CCWC), pp. 0071-0077, 2020.

Contact IEEE to Subscribe

References

References is not available for this document.