Maintained by Difan Deng and Marius Lindauer.
The following list considers papers related to neural architecture search. It is by no means complete. If you miss a paper on the list, please let us know.
Please note that although NAS methods steadily improve, the quality of empirical evaluations in this field are still lagging behind compared to other areas in machine learning, AI and optimization. We would therefore like to share some best practices for empirical evaluations of NAS methods, which we believe will facilitate sustained and measurable progress in the field. If you are interested in a teaser, please read our blog post or directly jump to our checklist.
Transformers have gained increasing popularity in different domains. For a comprehensive list of papers focusing on Neural Architecture Search for Transformer-Based spaces, the awesome-transformer-search repo is all you need.
2023
Kozarovytska, Polina; Kucherenko, Taras
Empirical comparison of hyperparameter optimization methods for neural networks Technical Manual
2023.
@manual{Kozarovytska23,
title = {Empirical comparison of hyperparameter optimization methods for neural networks},
author = {Polina Kozarovytska and Taras Kucherenko},
url = {https://www.researchgate.net/profile/Polina-Kozarovytska/publication/369597344_Empirical_comparison_of_hyperparameter_optimization_methods_for_neural_networks/links/6424430492cfd54f8439d45f/Empirical-comparison-of-hyperparameter-optimization-methods-for-neural-networks.pdf},
year = {2023},
date = {2023-03-17},
urldate = {2023-03-17},
keywords = {},
pubstate = {published},
tppubtype = {manual}
}
Yang, Jiechao; Liu, Yong; Xu, Hongteng
HOTNAS: Hierarchical Optimal Transport for Neural Architecture Search Technical Manual
2023.
@manual{yanghotnas-23a,
title = {HOTNAS: Hierarchical Optimal Transport for Neural Architecture Search},
author = {Jiechao Yang and Yong Liu and Hongteng Xu},
url = {https://gsai.ruc.edu.cn/uploads/20230325/e8599a4e5ddc3924b32b7248606fd06f.pdf},
year = {2023},
date = {2023-03-13},
urldate = {2023-03-13},
keywords = {},
pubstate = {published},
tppubtype = {manual}
}
Kaviani, Siavosh; Azar, Reza Salimpour
Does the new algorithm for Auto deep learning outperform the old algorithm in terms of accuracy and computational efficiency? Technical Report
2023.
@techreport{nokey,
title = {Does the new algorithm for Auto deep learning outperform the old algorithm in terms of accuracy and computational efficiency?},
author = {Siavosh Kaviani and Reza Salimpour Azar},
url = {https://www.researchgate.net/profile/Siavosh-Kaviani/publication/369094988_Does_the_new_algorithm_for_Auto_deep_learning_outperform_the_old_algorithm_in_terms_of_accuracy_and_computational_efficiency/links/6409cb7bbcd7982d8d6e77ae/Does-the-new-algorithm-for-Auto-deep-learning-outperform-the-old-algorithm-in-terms-of-accuracy-and-computational-efficiency.pdf},
year = {2023},
date = {2023-03-01},
urldate = {2023-03-01},
keywords = {},
pubstate = {published},
tppubtype = {techreport}
}
Tuli, S.; Jha, N. K.
EdgeTran: Device-Aware Co-Search Of Transformers for Efficient Inference on Mobile Edge Platforms Journal Article
In: IEEE Transactions on Mobile Computing, no. 01, pp. 1-18, 2023, ISSN: 1558-0660.
@article{10301516,
title = {EdgeTran: Device-Aware Co-Search Of Transformers for Efficient Inference on Mobile Edge Platforms},
author = {S. Tuli and N. K. Jha},
url = {https://www.computer.org/csdl/journal/tm/5555/01/10301516/1RFBMuHDM08},
doi = {10.1109/TMC.2023.3328287},
issn = {1558-0660},
year = {2023},
date = {2023-03-01},
urldate = {5555-10-01},
journal = {IEEE Transactions on Mobile Computing},
number = {01},
pages = {1-18},
publisher = {IEEE Computer Society},
address = {Los Alamitos, CA, USA},
abstract = {Automated design of efficient transformer models has recently attracted significant attention from industry and academia. However, most works only focus on certain metrics while searching for the best-performing transformer architecture. Furthermore, running traditional, complex, and large transformer models on low-compute edge platforms is a challenging problem. In this work, we propose a framework, called ProTran, to profile the hardware performance measures for a design space of transformer architectures and a diverse set of edge devices. We use this profiler in conjunction with the proposed co-search technique to obtain the best-performing models that have high accuracy on the given task and minimize latency, energy consumption, and peak power draw to enable edge deployment. We refer to our framework for co-optimizing accuracy and hardware performance measures as EdgeTran. It searches for the best transformer model and edge device pair. Finally, we propose GPTran, a multi-stage block-level grow-and-prune post-processing step that further improves accuracy in a hardware-aware manner. The obtained transformer model is 2.8× smaller and has a 0.8% higher GLUE score than the baseline (BERT-Base). Inference with it on the selected edge device enables 15.0% lower latency, 10.0× lower energy, and 10.8× lower peak power draw compared to an off-the-shelf GPU.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Sekkal, Mansouria; Benzina, Amina; badir Benkrelifa, Lahouari
Multi-Objective Evolutionary Algorithm based on NSGA-II for Neural Network Optimization Application to the Prediction of Severe Diseases Journal Article
In: Informatica, 2023.
@article{Sekkal-informatica23a,
title = {Multi-Objective Evolutionary Algorithm based on NSGA-II for Neural Network Optimization Application to the Prediction of Severe Diseases},
author = {Mansouria Sekkal and Amina Benzina and Lahouari badir Benkrelifa},
url = {https://www.informatica.si/index.php/informatica/article/view/5126},
year = {2023},
date = {2023-03-01},
urldate = {2023-03-01},
journal = {Informatica},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Z, Zhang; I, Joe
CAM-NAS: A Fast Model for Neural Architecture Search based on Class Activation Map Journal Article
In: Research Square, 2023.
@article{ZhangRS3,
title = {CAM-NAS: A Fast Model for Neural Architecture Search based on Class Activation Map},
author = {Zhang Z and Joe I
},
url = {https://europepmc.org/article/ppr/ppr623353},
year = {2023},
date = {2023-02-28},
journal = { Research Square},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Su, Lei; Zhou, Haoyi; Huang, Hua; Zhang, Wancai; Cao, Boyuan
Research on transformer internal defect detection based on large scale model Miscellaneous
2023.
@misc{surtid23,
title = {Research on transformer internal defect detection based on large scale model},
author = {Lei Su and Haoyi Zhou and Hua Huang and Wancai Zhang and Boyuan Cao},
url = {https://iopscience.iop.org/article/10.1088/1742-6596/2425/1/012042/pdf},
year = {2023},
date = {2023-02-22},
urldate = {2023-02-22},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Keisler, Julie; Talbi, El-Ghazali; Claudel, Sandra; Cabriel, Gilles
An algorithmic framework for the optimization of deep neural networks architectures and hyperparameters Technical Manual
2023.
@manual{KeislerHAL23,
title = {An algorithmic framework for the optimization of deep neural networks architectures and hyperparameters},
author = {Julie Keisler and El-Ghazali Talbi and Sandra Claudel and Gilles Cabriel},
url = {https://hal.science/hal-03982852/document},
year = {2023},
date = {2023-02-22},
urldate = {2023-02-22},
keywords = {},
pubstate = {published},
tppubtype = {manual}
}
Speckhard, Daniel T.; Misiunas, Karolis; Perel, Sagi; Zhu, Tenghui; Carlile, Simon; Slaney, Malcolm
Neural architecture search for energy-efficient always-on audio machine learning Journal Article
In: Neural Computing and Applications , 2023.
@article{SpeckhardNCA23,
title = {Neural architecture search for energy-efficient always-on audio machine learning},
author = {Daniel T. Speckhard and Karolis Misiunas and Sagi Perel and Tenghui Zhu and Simon Carlile and Malcolm Slaney
},
url = {https://link.springer.com/article/10.1007/s00521-023-08345-y},
year = {2023},
date = {2023-02-20},
urldate = {2023-02-20},
journal = {Neural Computing and Applications },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Laube, Kevin A.
Improving the automated search of neural network architectures PhD Thesis
2023.
@phdthesis{laube-phd2023,
title = {Improving the automated search of neural network architectures},
author = {Kevin A. Laube},
url = {https://tobias-lib.ub.uni-tuebingen.de/xmlui/bitstream/handle/10900/138640/Dissertation.pdf?sequence=2&isAllowed=y},
year = {2023},
date = {2023-02-16},
urldate = {2023-02-16},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
Du, Yangyi; Zhou, Xiaojun; Huang, Tingwen; Yang, Chunhua
A hierarchical evolution of neural architecture search method based on state transition algorithm Bachelor Thesis
2023.
@bachelorthesis{Du-ijmlc23,
title = {A hierarchical evolution of neural architecture search method based on state transition algorithm},
author = {
Yangyi Du and Xiaojun Zhou and Tingwen Huang and Chunhua Yang
},
url = {https://link.springer.com/article/10.1007/s13042-023-01794-w},
year = {2023},
date = {2023-02-13},
urldate = {2023-02-13},
journal = { International Journal of Machine Learning and Cybernetics },
keywords = {},
pubstate = {published},
tppubtype = {bachelorthesis}
}
Mohammadrezaei, Parsa; Aminan, Mohammad; Soltanian, Mohammad; Borna, Keivan
Improving CNN-based solutions for emotion recognition using evolutionary algorithms Journal Article
In: Results in Applied Mathematics, 2023.
@article{MohammadrezaeiRAM22,
title = {Improving CNN-based solutions for emotion recognition using evolutionary algorithms},
author = {Parsa Mohammadrezaei and Mohammad Aminan and Mohammad Soltanian and Keivan Borna },
url = {https://www.researchgate.net/profile/Mohammad-Aminan/publication/368576737_Improving_CNN-based_solutions_for_emotion_recognition_using_evolutionary_algorithms/links/63ef6d7d51d7af0540325e48/Improving-CNN-based-solutions-for-emotion-recognition-using-evolutionary-algorithms.pdf},
year = {2023},
date = {2023-02-13},
urldate = {2023-02-13},
journal = {Results in Applied Mathematics},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Stephen, Okeke; Sain, Mangal
In: Developments in Optimization Algorithms for Smart Healthcare 2022, 2023.
@article{stephen-doafsh22,
title = {Using Deep Learning with Bayesian–Gaussian Inspired Convolutional Neural Architectural Search for Cancer Recognition and Classification from Histopathological Image Frames},
author = {Okeke Stephen and Mangal Sain},
url = {https://www.hindawi.com/journals/jhe/2023/4597445/},
year = {2023},
date = {2023-02-09},
urldate = {2023-02-09},
journal = {Developments in Optimization Algorithms for Smart Healthcare 2022},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Evolving Deep Neural Networks with Explanations for Image Classification PhD Thesis
2023.
@phdthesis{wang-phd23,
title = {Evolving Deep Neural Networks with Explanations for Image Classification},
url = {https://s3-ap-southeast-2.amazonaws.com/pstorage-wellington-7594921145/38986799/thesis_access.pdf?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIA3OGA3B5WBO5PUAXV/20230208/ap-southeast-2/s3/aws4_request&X-Amz-Date=20230208T183516Z&X-Amz-Expires=10&X-Amz-SignedHeaders=host&X-Amz-Signature=598d7d8640153b25c41deadcb126a9cf1656dffaeb6797dc58f70e16638eecc2},
year = {2023},
date = {2023-02-03},
urldate = {2023-02-03},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
Zawad, Syed
Towards Scalable, Private, and Practical Deep Learning PhD Thesis
2023.
@phdthesis{zawad-phd23,
title = {Towards Scalable, Private, and Practical Deep Learning},
author = {Syed Zawad},
url = {https://scholarworks.unr.edu/bitstream/handle/11714/8356/Zawad_unr_0139D_13906.pdf?sequence=1&isAllowed=y},
year = {2023},
date = {2023-02-03},
urldate = {2023-02-03},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
Zheng, Xiawu; Yang, Chenyi; Zhang, Shaokun; Wang, Yan; Zhang, Baochang; Wu, Yongjian; Wu, Yunsheng; Shao, Ling; Ji, Rongrong
DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution Pruning Journal Article
In: International Journal of Computer Vision , 2023.
@article{Zhengijcv23,
title = {DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution Pruning},
author = {Xiawu Zheng and Chenyi Yang and Shaokun Zhang and Yan Wang and Baochang Zhang and Yongjian Wu and Yunsheng Wu and Ling Shao and Rongrong Ji
},
url = {https://link.springer.com/article/10.1007/s11263-023-01753-6},
year = {2023},
date = {2023-02-03},
urldate = {2023-02-03},
journal = {International Journal of Computer Vision },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Islama, Usman; Mahuma, Rabbia; AlSalman, AbdulMalik
Facial Emotions Detection using an Efficient Neural Architecture Search Network Technical Report
2023.
@techreport{IslamaRS23,
title = {Facial Emotions Detection using an Efficient Neural Architecture Search Network},
author = {Usman Islama and Rabbia Mahuma and AbdulMalik AlSalman},
url = {https://assets.researchsquare.com/files/rs-2526836/v1_covered.pdf?c=1675443603},
year = {2023},
date = {2023-02-03},
urldate = {2023-02-03},
keywords = {},
pubstate = {published},
tppubtype = {techreport}
}
Wang, Q.; Zhang, S.
DGL: Device Generic Latency Model for Neural Architecture Search on Mobile Devices Journal Article
In: IEEE Transactions on Mobile Computing, no. 01, pp. 1-14, 2023, ISSN: 1558-0660.
@article{10042973,
title = {DGL: Device Generic Latency Model for Neural Architecture Search on Mobile Devices},
author = {Q. Wang and S. Zhang},
url = {https://www.computer.org/csdl/journal/tm/5555/01/10042973/1KJs8PnAasw},
doi = {10.1109/TMC.2023.3244170},
issn = {1558-0660},
year = {2023},
date = {2023-02-01},
urldate = {5555-02-01},
journal = {IEEE Transactions on Mobile Computing},
number = {01},
pages = {1-14},
publisher = {IEEE Computer Society},
address = {Los Alamitos, CA, USA},
abstract = {The low-cost Neural Architecture Search (NAS) for lightweight networks working on massive mobile devices is essential for fast-developing ICT technology. Current NAS work can not search on unseen devices without latency sampling, which is a big obstacle to the implementation of NAS on mobile devices. In this paper, we overcome this challenge by proposing the Device Generic Latency (DGL) model. By absorbing processor modeling technology, the proposed DGL formula maps the parameters in the interval theory to the seven static configuration parameters of the device. And to make the formula more practical, we refine it to low-cost form by decreasing the number of configuration parameters to four. Then based on this formula, the DGL model is proposed which introduces the network parameters predictor and accuracy predictor to work with the DGL formula to predict the network latency. We propose the DGL-based NAS framework to enable fast searches without latency sampling. Extensive experiments results validate that the DGL model can achieve more accurate latency predictions than existing NAS latency predictors on unseen mobile devices. When configured with current state-of-the-art predictors, DGL-based NAS can search for architectures with higher accuracy that meet the latency limit than other NAS implementations, while using less training time and prediction time. Our work shed light on how to adopt domain knowledge into NAS topic and play important role in low-cost NAS on mobile devices.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wei, Lanning; Zhao, Huan; He, Zhiqiang; Yao, Quanming
Neural Architecture Search for GNN-Based Graph Classification Journal Article
In: ACM Trans. Inf. Syst., 2023, ISSN: 1046-8188, (Just Accepted).
@article{10.1145/3584945,
title = {Neural Architecture Search for GNN-Based Graph Classification},
author = {Lanning Wei and Huan Zhao and Zhiqiang He and Quanming Yao},
url = {https://doi.org/10.1145/3584945},
doi = {10.1145/3584945},
issn = {1046-8188},
year = {2023},
date = {2023-02-01},
urldate = {2023-02-01},
journal = {ACM Trans. Inf. Syst.},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {Graph classification is an important problem with applications across many domains, for which the graph neural networks (GNNs) have been state-of-the-art (SOTA) methods. In the literature, to adopt GNNs for the graph classification task, there are two groups of methods: global pooling and hierarchical pooling. The global pooling methods obtain the graph representation vectors by globally pooling all the node embeddings together at the end of several GNN layers, while the hierarchical pooling methods provide one extra pooling operation between the GNN layers to extract the hierarchical information and improve the graph representations. Both global and hierarchical pooling methods are effective in different scenarios. Due to highly diverse applications, it is challenging to design data-specific pooling methods with human expertise. To address this problem, we propose PAS (Pooling Architecture Search) to design adaptive pooling architectures by using the neural architecture search (NAS). To enable the search space design, we propose a unified pooling framework consisting of four modules: Aggregation, Pooling, Readout, and Merge. Two variants PAS-G and PAS-NE are provided to design the pooling operations in different scales. A set of candidate operations are designed in the search space on top of this framework, and then existing human-designed pooling methods, including global and hierarchical ones, can be incorporated. To enable efficient search, a coarsening strategy is developed to continuously relax the search space, and then a differentiable search method can be adopted. We conduct extensive experiments on six real-world datasets, including the large-scale datasets MR and ogbg-molhiv. Experimental results in this paper demonstrate the effectiveness and efficiency of the proposed PAS in designing the pooling architectures for graph classification. Besides, the Top-1 performance on two Open Graph Benchmark (OGB) datasets further indicates the utility of PAS when facing diverse realistic data. The implementation of PAS is available at: https://github.com/AutoML-Research/PAS.},
note = {Just Accepted},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Ci, Yuanzheng
Efficient Methods for the Design and Training of Neural Networks PhD Thesis
2023.
@phdthesis{Ci-phd23,
title = {Efficient Methods for the Design and Training of Neural Networks},
author = {Yuanzheng Ci},
url = {https://ses.library.usyd.edu.au/bitstream/handle/2123/31183/Doctorial_Thesis_v2.0.pdf?sequence=1&isAllowed=y},
year = {2023},
date = {2023-02-01},
urldate = {2023-02-01},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
(Ed.)
Neural Architecture Search for Wide Spectrum Adversarial Robustness Collection
2023.
@collection{ChengAAAI23,
title = {Neural Architecture Search for Wide Spectrum Adversarial Robustness},
author = {Zhi Cheng and Yanxi Li and Minjing Dong and Xiu Su and Shan You and Chang Xu},
url = {https://ojs.aaai.org/index.php/AAAI/article/download/25118/24890},
year = {2023},
date = {2023-02-01},
urldate = {2023-02-01},
booktitle = {The Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI-23)},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
Fang, Yuchu; Li, Wenzhong; Zeng, Yao; Zheng, Yang; Hu, Zheng; Lu, Sanglu
PatchNAS: Repairing DNNs in Deployment with Patched Network Architecture Search Conference
https://ojs.aaai.org/index.php/AAAI/article/download/26730/26502, 2023.
@conference{Fang-aaai23,
title = {PatchNAS: Repairing DNNs in Deployment with Patched Network Architecture Search},
author = {Yuchu Fang and Wenzhong Li and Yao Zeng and Yang Zheng and Zheng Hu and Sanglu Lu},
url = {https://ojs.aaai.org/index.php/AAAI/article/download/26730/26502},
year = {2023},
date = {2023-02-01},
urldate = {2023-02-01},
booktitle = {https://ojs.aaai.org/index.php/AAAI/article/download/26730/26502},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Jia, Liang; Tian, Ye; Zhang, Junguo
Neural architecture search based on packed samples for identifying animals in camera trap images Journal Article
In: Neural Computing and Applications (2023), 2023.
@article{JiaNCA23,
title = {Neural architecture search based on packed samples for identifying animals in camera trap images},
author = {Liang Jia and Ye Tian and Junguo Zhang },
url = {https://link.springer.com/article/10.1007/s00521-023-08247-z},
year = {2023},
date = {2023-01-29},
urldate = {2023-01-29},
journal = {Neural Computing and Applications (2023)},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Tchetchenian, Ari; Zhu, Yanming; Zhang, Fan; O’Donnell, Lauren J.; Song, Yang; Meijering, Erik
A comparison of manual and automated neural architecture search for white matter tract segmentation Journal Article
In: Scientific Reports , 2023.
@article{TchetchenianSR23,
title = {A comparison of manual and automated neural architecture search for white matter tract segmentation},
author = {
Ari Tchetchenian and Yanming Zhu and Fan Zhang and Lauren J. O’Donnell and Yang Song and Erik Meijering
},
url = {https://www.nature.com/articles/s41598-023-28210-1},
year = {2023},
date = {2023-01-28},
urldate = {2023-01-28},
journal = {Scientific Reports },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Yue, Zhixiong; Zhang, Yu; Liang, Jie
Learning Conflict-Noticed Architecture for Multi-Task Learning Miscellaneous
2023.
@misc{Zue23,
title = {Learning Conflict-Noticed Architecture for Multi-Task Learning},
author = {Zhixiong Yue and Yu Zhang and Jie Liang},
url = {https://yuezhixiong.github.io/Papers/CoNAL.pdf},
year = {2023},
date = {2023-01-26},
urldate = {2023-01-26},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Chang, Chen-Chia; Pan, Jingyu; Xie, Zhiyao; Li, Yaguang; Lin, Yishuang; Hu, Jiang; Chen, Yiran
Fully Automated Machine Learning Model Development for Analog Placement Quality Prediction Proceedings Article
In: 2023 Asia and South Pacific Design Automation Conference (ASP-DAC), 2023.
@inproceedings{ChangASPDAC23,
title = {Fully Automated Machine Learning Model Development for Analog Placement Quality Prediction},
author = {Chen-Chia Chang and Jingyu Pan and Zhiyao Xie and Yaguang Li and Yishuang Lin and Jiang Hu and Yiran Chen},
url = {https://zhiyaoxie.github.io/files/ASPDAC23_NAS_Analog.pdf},
year = {2023},
date = {2023-01-19},
urldate = {2023-01-19},
booktitle = {2023 Asia and South Pacific Design Automation Conference (ASP-DAC)},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Zhang, Jian; Gong, Xuan; Liu, YuXiao; Wang, Wei; Wang, Lei; Zhang, BaoChang
Bandit neural architecture search based on performance evaluation for operation selection Journal Article
In: Science China Technological Sciences, 2023.
@article{ZhangSCTS23,
title = {Bandit neural architecture search based on performance evaluation for operation selection},
author = {Jian Zhang and Xuan Gong and YuXiao Liu and Wei Wang and Lei Wang and BaoChang Zhang
},
url = {https://link.springer.com/article/10.1007/s11431-022-2197-y},
year = {2023},
date = {2023-01-16},
urldate = {2023-01-16},
journal = {Science China Technological Sciences},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Asadi, Mehdi; Poursalim, Fatemeh; Loni, Mohammad; Daneshtalab, Masoud; Sjödin, Mikael; Gharehbaghi, Arash
Accurate Detection of Paroxysmal Atrial Fibrillation with Certified-GAN and Neural Architecture Search Technical Manual
2023.
@manual{nokey,
title = {Accurate Detection of Paroxysmal Atrial Fibrillation with Certified-GAN and Neural Architecture Search},
author = {Mehdi Asadi and Fatemeh Poursalim and Mohammad Loni and Masoud Daneshtalab and Mikael Sjödin and Arash Gharehbaghi},
url = {https://assets.researchsquare.com/files/rs-2485416/v1/03c1c5a579b494a39449a4ee.pdf?c=1674177185},
year = {2023},
date = {2023-01-16},
urldate = {2023-01-16},
keywords = {},
pubstate = {published},
tppubtype = {manual}
}
Zhang, Jian; Gong, Xuan; Liu, YuXiao; Wang, Wei; Wang, Lei; Zhang, BaoChang
Bandit neural architecture search based on performance evaluation for operation selection Journal Article
In: Science China Technological Sciences 2023, 2023.
@article{ZhangSCTS23b,
title = {Bandit neural architecture search based on performance evaluation for operation selection},
author = {
Jian Zhang and Xuan Gong and YuXiao Liu and Wei Wang and Lei Wang and BaoChang Zhang
},
url = {https://link.springer.com/article/10.1007/s11431-022-2197-y},
year = {2023},
date = {2023-01-16},
urldate = {2023-01-16},
journal = {Science China Technological Sciences 2023},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Franchini, Giorgia; Ruggiero, Valeria; Porta, Federica; Zanni, Luca
Neural architecture search via standard machine learning methodologies Journal Article
In: Mathematics in Engineering, 2023.
@article{Franchini2023,
title = {Neural architecture search via standard machine learning methodologies},
author = {Giorgia Franchini and Valeria Ruggiero and Federica Porta and Luca Zanni},
url = {https://www.aimspress.com/article/doi/10.3934/mine.2023012?viewType=HTML},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Mathematics in Engineering},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wang, Ziyan; Qi, Feng; Zou, Liming
Continuous Evolution for Efficient Neural Architecture Search Based on Improved NSGA-III Algorithm Proceedings Article
In: Sun, Jiande; Wang, Yue; Huo, Mengyao; Xu, Lexi (Ed.): Signal and Information Processing, Networking and Computers, pp. 979–986, Springer Nature Singapore, Singapore, 2023, ISBN: 978-981-19-3387-5.
@inproceedings{10.1007/978-981-19-3387-5_117,
title = {Continuous Evolution for Efficient Neural Architecture Search Based on Improved NSGA-III Algorithm},
author = {Ziyan Wang and Feng Qi and Liming Zou},
editor = {Jiande Sun and Yue Wang and Mengyao Huo and Lexi Xu},
isbn = {978-981-19-3387-5},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Signal and Information Processing, Networking and Computers},
pages = {979--986},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Improved Continuous Evolution for Efficient Neural Architecture Search method (I-CARS) is proposed to solve multi-objective optimization problems (MOPs). Seeking to improve the convergence and search accuracy, two modifications of the non-dominated sorting genetic algorithm based on reference-point strategy (NSGA-III) were made to replace pNSGA-III in CARS, including the penalty-based boundary intersection distance (PBI distance) and the selection-and-elimination operators. First, the perpendicular distance from solutions to reference lines was replaced by the PBI distance in the offspring selection stage phase, which can add the convergence information. Second, individuals in the population were selected or eliminated by niche count and PBI distance. Better performing individuals were selected to become the next generation, while poorly performing individuals were eliminated. The convergence and diversity of the population can be balanced by adding the selection-and-elimination operator. Experiments were conducted on Vega pipeline, and I-CARS achieves 3.41% test error on CIFAR10, the results indicated that the accuracy and convergence of I-CARS are enhanced compared to CARS.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Rajarajeswari, S.; Patil, Annapurna P.; Madhyastha, Aditya; Jaitly, Akshat; Jha, Himangshu Shekhar; Bhave, Sahil Rajesh; Das, Mayukh; Pradeep, N. S.
Design and Develop Hardware Aware DNN for Faster Inference Proceedings Article
In: Ärai, Kohei" (Ed.): Intelligent Systems and Applications, pp. 309–318, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-16075-2.
@inproceedings{10.1007/978-3-031-16075-2_21,
title = {Design and Develop Hardware Aware DNN for Faster Inference},
author = {S. Rajarajeswari and Annapurna P. Patil and Aditya Madhyastha and Akshat Jaitly and Himangshu Shekhar Jha and Sahil Rajesh Bhave and Mayukh Das and N. S. Pradeep},
editor = {Kohei" Ärai},
isbn = {978-3-031-16075-2},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Intelligent Systems and Applications},
pages = {309--318},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {Ön many small-scale devices, advanced learning models have become standard. The necessity of the hour is to reduce the amount of time required for inference. This study describes a pipeline for automating Deep Neural Network customization and reducing neural network inference time. This paper presents a hardware-aware methodology in the form of a sequential pipeline for shrinking the size of deep neural networks. MorphNet is used at the pipeline's core to iteratively decrease and enlarge a network. Upon the activation of layers, a resource-weighted sparsifying regularizer is used to identify and prune inefficient neurons, and all layers are then expanded using a uniform multiplicative factor. This is followed by fusion, a technique for combining the frozen batch normalization layer with the preceding convolution layer. Finally, the DNN is retrained after customization using a Knowledge Distillation approach to maintain model accuracy performance. The approach shows promising initial results on MobileNetv1 and ResNet50 architectures."},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Liang, Jingkang; Liao, Yixiao; Li, Weihua
Differentiable Architecture Searched Network with Tree-Structured Parzen Estimators for Rotating Machinery Fault Diagnosis Proceedings Article
In: Zhang, Hao; Feng, Guojin; Wang, Hongjun; Gu, Fengshou; Sinha, Jyoti K. (Ed.): Proceedings of IncoME-VI and TEPEN 2021, pp. 959–970, Springer International Publishing, Cham, 2023, ISBN: 978-3-030-99075-6.
@inproceedings{10.1007/978-3-030-99075-6_77,
title = {Differentiable Architecture Searched Network with Tree-Structured Parzen Estimators for Rotating Machinery Fault Diagnosis},
author = {Jingkang Liang and Yixiao Liao and Weihua Li},
editor = {Hao Zhang and Guojin Feng and Hongjun Wang and Fengshou Gu and Jyoti K. Sinha},
isbn = {978-3-030-99075-6},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Proceedings of IncoME-VI and TEPEN 2021},
pages = {959--970},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {Deep learning is widely used in the field of rotating machinery fault diagnosis. However, manually designing the neural network structure and adjusting the hyperparameters for specific fault diagnosis task are complex and requires a lot of expert knowledge. Aiming at these problems, Differentiable Architecture Searched Network with Tree-Structured Parzen Estimators (DASNT) is proposed for fault diagnosis. Differentiable Architecture Search (DARTS) is utilized to automatically search network structure for specific fault diagnosis task. Tree-Structured Parzen Estimators (TPE) is utilized to optimize the hyperparameters of the network searched by DARTS, which can further improve the fault diagnosis accuracy. The results of comparison experiments indicate that the network architecture and hyperparameters optimized by DASNT can achieve superior fault diagnosis performance.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Liu, Yang; Liang, Xinle; Luo, Jiahuan; He, Yuanqin; Chen, Tianjian; Yao, Quanming; Yang, Qiang
Cross-Silo Federated Neural Architecture Search for Heterogeneous and Cooperative Systems Book Chapter
In: Razavi-Far, Roozbeh; Wang, Boyu; Taylor, Matthew E.; Yang, Qiang (Ed.): Federated and Transfer Learning, pp. 57–86, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-11748-0.
@inbook{Liu2023,
title = {Cross-Silo Federated Neural Architecture Search for Heterogeneous and Cooperative Systems},
author = {Yang Liu and Xinle Liang and Jiahuan Luo and Yuanqin He and Tianjian Chen and Quanming Yao and Qiang Yang},
editor = {Roozbeh Razavi-Far and Boyu Wang and Matthew E. Taylor and Qiang Yang},
url = {https://doi.org/10.1007/978-3-031-11748-0_4},
doi = {10.1007/978-3-031-11748-0_4},
isbn = {978-3-031-11748-0},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Federated and Transfer Learning},
pages = {57--86},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {In many cooperative systems (i.e. autonomous vehicles, robotics, hospital networks), data are privately and heterogeneously distributed among devices with various computational constraints, and no party has a global view of data or device distribution. Federated Neural Architecture Search (FedNAS) was previously proposed to adapt Neural Architecture Search (NAS) into Federated Learning (FL) to provide both privacy and model performance to such uninspectable and heterogeneous systems. However, these approaches mostly apply to scenarios where parties share the same data attributes and comparable computation resources. In this chapter, we present Self-supervised Vertical Federated Neural Architecture Search (SS-VFNAS) for automating FL where participants have heterogeneous data and resource constraints, a common cross-silo scenario. SS-VFNAS not only simultaneously optimizes all parties' model architecture and parameters for the best global performance under a vertical FL (VFL) framework using only a small set of aligned and labeled data, but also preserves each party's local optimal model architecture under a self-supervised NAS framework. We demonstrate that SS-VFNAS is a promising framework of superior performance, communication efficiency and privacy, and is capable of generating high-performance and highly-transferable heterogeneous architectures with only limited overlapping samples, providing practical solutions for designing collaborative systems with both limited data and resource constraints.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Deng, TianJin; Wu, Jia
Efficient graph neural architecture search using Monte Carlo Tree search and prediction network Journal Article
In: Expert Systems with Applications, vol. 213, pp. 118916, 2023, ISSN: 0957-4174.
@article{DENG2023118916,
title = {Efficient graph neural architecture search using Monte Carlo Tree search and prediction network},
author = {TianJin Deng and Jia Wu},
url = {https://www.sciencedirect.com/science/article/pii/S0957417422019340},
doi = {https://doi.org/10.1016/j.eswa.2022.118916},
issn = {0957-4174},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Expert Systems with Applications},
volume = {213},
pages = {118916},
abstract = {Graph Neural Networks (GNNs) have emerged recently as a powerful way of dealing with non-Euclidean data on graphs, such as social networks and citation networks. Despite their success, obtaining optimal graph neural networks requires immense manual work and domain knowledge. Inspired by the strong searching capability of neural architecture search in CNN, a few attempts automatically search optimal GNNs that rival the best human-invented architectures. However, existing Graph Neural Architecture Search (GNAS) approaches face two challenges: (1) Sampling GNNs across the entire search space results in low search efficiency, particularly in large search spaces. (2) It is pretty costly to evaluate GNNs by training architectures from scratch. To overcome these challenges, this paper proposes an Efficient Graph Neural Architecture Search (EGNAS) method based on Monte Carlo Tree Search (MCTS) and a prediction network. Specifically, EGNAS first uses MCTS to recursively partition the entire search space into good or bad search regions. Then, the reinforcement learning-based search strategy (also called the agent) is applied to sample GNNs in those good search regions, which prevents overly exploring complex architectures and bad-performance regions, thus improving sampling efficiency. To reduce the evaluation cost, we use a prediction network to estimate the performance of GNNs. We alternately use ground-truth accuracy (by training GNNs from scratch) and prediction accuracy (by the prediction network) to update the search strategy to avoid inaccuracies caused by long-term use of the prediction network. Furthermore, to improve the training efficiency and stability, the agent is trained by a variant of Proximal Policy Optimization. Experiments show that EGNAS can search for better GNNs in the promising search region in a shorter search time, with an accuracy of 83.5%, 73.3%, 79.6%, and 94.5% on Cora, Citeseer, Pubmed, and Photo datasets, respectively In particular, compared to the most popular GNAS algorithm, our EGNAS-NP without using the prediction network achieves an accuracy of 83.6% on Cora, 73.5% on Citeseer, 79.9% on Pubmed, and 94.6% on Photo, with a relative improvement of 0.6%, 0.2%, 0.7%, and 0.6%.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Souquet, Léo; Shvai, Nadiya; Llanza, Arcadi; Nakib, Amir
Convolutional neural network architecture search based on fractal decomposition optimization algorithm Journal Article
In: Expert Systems with Applications, vol. 213, pp. 118947, 2023, ISSN: 0957-4174.
@article{SOUQUET2023118947,
title = {Convolutional neural network architecture search based on fractal decomposition optimization algorithm},
author = {Léo Souquet and Nadiya Shvai and Arcadi Llanza and Amir Nakib},
url = {https://www.sciencedirect.com/science/article/pii/S0957417422019650},
doi = {https://doi.org/10.1016/j.eswa.2022.118947},
issn = {0957-4174},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Expert Systems with Applications},
volume = {213},
pages = {118947},
abstract = {This paper presents a new approach to design the architecture and optimize the hyperparameters of a deep convolutional neural network (CNN) via of the Fractal Decomposition Algorithm (FDA). This optimization algorithm was recently proposed to solve continuous optimization problems. It is based on a geometric fractal decomposition that divides the search space while searching for the best solution possible. As FDA is effective in single-objective optimization, in this work we aim to prove that it can also be successfully applied to fine-tuning deep neural network architectures. Moreover, a new formulation based on bi-level optimization is proposed to separate the architecture search composed of discrete parameters from hyperparameters’ optimization. This is motivated by the fact that automating the construction of deep neural architecture has been an important focus over recent years as manual construction is considerably time-consuming, error-prone, and requires in-depth knowledge. To solve the bi-level problem thus formulated, a random search is performed aiming to create a set of candidate architectures. Then, the best ones are finetuned using FDA. CIFAR-10 and CIFAR-100 benchmarks were used to evaluate the performance of the proposed approach. The results obtained are among the state of the art in the corresponding class of networks (low number of parameters and chained-structured CNN architectures). The results are emphasized by the fact that the whole process was performed using low computing power with only 3 NVIDIA V100 GPUs. The source code is available at https://github.com/alc1218/Convolutional-Neural-Network-Architecture-Search-Based-on-Fractal-Decomposition-Optimization.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wen, Yingpeng; Yu, Weijiang; Li, Dongsheng; Du, Jiangsu; Huang, Dan; Xiao, Nong
CosNAS: Enhancing estimation on cosmological parameters via neural architecture search Journal Article
In: New Astronomy, vol. 99, pp. 101955, 2023, ISSN: 1384-1076.
@article{WEN2023101955,
title = {CosNAS: Enhancing estimation on cosmological parameters via neural architecture search},
author = {Yingpeng Wen and Weijiang Yu and Dongsheng Li and Jiangsu Du and Dan Huang and Nong Xiao},
url = {https://www.sciencedirect.com/science/article/pii/S1384107622001373},
doi = {https://doi.org/10.1016/j.newast.2022.101955},
issn = {1384-1076},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {New Astronomy},
volume = {99},
pages = {101955},
abstract = {A great challenge of cosmology is estimating the cosmological parameters of the universe. With the development of deep learning, scientists adopt 3D deep neural networks to estimate cosmological parameters from the large-scale dark matter distribution of the universe, but these methods are time-consuming to design and train neural networks. While neural architecture search is an emerging approach to estimate cosmological parameters with its capability of automatically designing neural networks, the 3D operations on a 3D dataset prohibit the usage of traditional neural architecture search methods, due to its overwhelming time and memory consumption. To tackle these issues, we propose an efficient method, CosNAS, that can automatically design neural networks with 2D operations to estimate the cosmological parameters. In addition, processing 3D data with 2D operations will inevitably cause the loss of spatial information, thus we propose an efficient SABlock to retain more 3D spatial information. We also customize a space-focused search space to focus on important information in the dark matter distribution. The experimental results indicate that our estimation of the cosmological parameters Ω, σ and n, can be applied to large-scale 3D dark matter distribution and speedup the network search by 800x. The average relative errors of cosmological parameter estimations are (0.00163, 0.00065, 0.00080), significantly decreasing the average error of estimation by 85.5% compared to previous work.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Li, Chuanyou; Zhang, Kun; Li, Yifan; Shang, Jiangwei; Zhang, Xinyue; Qian, Lei
ANNA: Accelerating Neural Network Accelerator through software-hardware co-design for vertical applications in edge systems Journal Article
In: Future Generation Computer Systems, vol. 140, pp. 91-103, 2023, ISSN: 0167-739X.
@article{LI202391,
title = {ANNA: Accelerating Neural Network Accelerator through software-hardware co-design for vertical applications in edge systems},
author = {Chuanyou Li and Kun Zhang and Yifan Li and Jiangwei Shang and Xinyue Zhang and Lei Qian},
url = {https://www.sciencedirect.com/science/article/pii/S0167739X22003168},
doi = {https://doi.org/10.1016/j.future.2022.10.001},
issn = {0167-739X},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Future Generation Computer Systems},
volume = {140},
pages = {91-103},
abstract = {In promising edge systems, AI algorithms and their hardware implementations are often joint optimized as integrated solutions to solve end-to-end design problems. Joint optimization depends on a delicate co-design of software and hardware. According to our knowledge, current co-design methodologies are still coarse-grained. In this paper, we proposed ANNA: Accelerating Neural Network Accelerator through a novel software-hardware co-design methodology. ANNA is a framework composed of three components: ANNA-NAS (Neural Architecture Search), ANNA-ARCH (hardware ARCHitecture) and ANNA-PERF (PERFormance optimizer & evaluator). ANNA-NAS adopts a cell-wise structure and is designed to be hardware aware. It aims at generating a neural network having high inference accuracy and low inference latency. To avoid tremendous time costs, ANNA-NAS synthetically uses differentiable architecture search and early stopping techniques. ANNA-ARCH starts to be designed as long as the architecture search space is defined. Based on the cell-wise structure, ANNA-ARCH specifies its main body which includes Convolution units, Activation Router and Buffer Pool. To well support different neural networks that could be generated by ANNA-NAS, the detailed part of ANNA-ARCH is configurable. ANNA-PERF harmonizes the co-design of ANNA-NAS and ANNA-ARCH. It takes a neural network and a hardware architecture as inputs. After optimizing the mapping strategy between the neural network and hardware accelerator, it feeds back a cycle-accurate latency to ANNA-NAS. Aiming at image classification, we carried out the experiments on ImageNet. Experimental results demonstrate that without loss of much inference accuracy, ANNA wins a significant low inference latency through a harmonious software and hardware co-design.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Zharikov, Ilia; Krivorotov, Ivan; Maximov, Egor; Korviakov, Vladimir; Letunovskiy, Alexey
Ä Review of One-Shot Neural Architecture Search Methods Proceedings Article
In: Kryzhanovsky, Boris; Dunin-Barkowski, Witali; Redko, Vladimir; Tiumentsev, Yury (Ed.): Ädvances in Neural Computation, Machine Learning, and Cognitive Research VI", pp. 130–147, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-19032-2.
@inproceedings{10.1007/978-3-031-19032-2_14,
title = {Ä Review of One-Shot Neural Architecture Search Methods},
author = {Ilia Zharikov and Ivan Krivorotov and Egor Maximov and Vladimir Korviakov and Alexey Letunovskiy},
editor = {Boris Kryzhanovsky and Witali Dunin-Barkowski and Vladimir Redko and Yury Tiumentsev},
isbn = {978-3-031-19032-2},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Ädvances in Neural Computation, Machine Learning, and Cognitive Research VI"},
pages = {130--147},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {Neural network architecture design is a challenging and computational expensive problem. For this reason training a one-shot model becomes very popular way to obtain several architectures or find the best according to different requirements without retraining. In this paper we summarize the existing one-shot NAS methods, highlight base concepts and compare considered methods in terms of accuracy, number of needed for training GPU hours and ranking quality.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Kolganov, Pavel A.; Tiumentsev, Yury V.
Än Attempt to Formalize the Formulation of the Network Architecture Search Problem for Convolutional Neural Networks Proceedings Article
In: Kryzhanovsky, Boris; Dunin-Barkowski, Witali; Redko, Vladimir; Tiumentsev, Yury (Ed.): Ädvances in Neural Computation, Machine Learning, and Cognitive Research VI", pp. 550–556, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-19032-2.
@inproceedings{10.1007/978-3-031-19032-2_55,
title = {Än Attempt to Formalize the Formulation of the Network Architecture Search Problem for Convolutional Neural Networks},
author = {Pavel A. Kolganov and Yury V. Tiumentsev},
editor = {Boris Kryzhanovsky and Witali Dunin-Barkowski and Vladimir Redko and Yury Tiumentsev},
isbn = {978-3-031-19032-2},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Ädvances in Neural Computation, Machine Learning, and Cognitive Research VI"},
pages = {550--556},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {The paper deals with the problem of searching for a neural network architecture. The paper presents a mathematical formulation of the problem of searching a neural network model, optimal from the point of view of a predefined criterion. The analysis of components of this problem is given. Some difficulties encountered by researchers when solving the NAS problem are described. A computational experiment is conducted, which consists in the search of a neural network architecture on the MNIST dataset.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Graham-Knight, John Brandon; Bond, Corey; Najjaran, Homayoun; Lucet, Yves; Lasserre, Patricia
Predicting and explaining performance and diversity of neural network architecture for semantic segmentation Journal Article
In: Expert Systems with Applications, vol. 214, pp. 119101, 2023, ISSN: 0957-4174.
@article{GRAHAMKNIGHT2023119101,
title = {Predicting and explaining performance and diversity of neural network architecture for semantic segmentation},
author = {John Brandon Graham-Knight and Corey Bond and Homayoun Najjaran and Yves Lucet and Patricia Lasserre},
url = {https://www.sciencedirect.com/science/article/pii/S0957417422021194},
doi = {https://doi.org/10.1016/j.eswa.2022.119101},
issn = {0957-4174},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Expert Systems with Applications},
volume = {214},
pages = {119101},
abstract = {This paper proposes searching for network architectures which achieve similar performance while promoting diversity, in order to facilitate ensembling. We explain prediction performance and diversity of various network sizes and activation functions applied to semantic segmentation of the CityScapes dataset. We show that both performance and diversity can be predicted from neural network architecture using explainable boosting machines. A grid search of 144 models is performed, and many of the models exhibit no significant difference in mean performance within a 95% confidence interval. Notably, we find the best performing models have varied network architecture parameters. The explanations for performance largely agree with the accepted wisdom of the machine learning community, which shows that the method is extracting information of value. We find that diversity between models can be achieved by varying network size. Moreover, homogeneous network sizes generally show positive correlation in predictions, and larger models tend to converge to similar solutions. These explanations provide a better understanding of the effects of network parameters to deep learning practitioners; they could also be used in place of naïve search methods or a model pool to inform growing an ensemble.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Sun, Yanan; Yen, Gary G.; Zhang, Mengjie
End-to-End Performance Predictors Book Chapter
In: Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances, pp. 237–255, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-16868-0.
@inbook{Sun2023,
title = {End-to-End Performance Predictors},
author = {Yanan Sun and Gary G. Yen and Mengjie Zhang},
url = {https://doi.org/10.1007/978-3-031-16868-0_14},
doi = {10.1007/978-3-031-16868-0_14},
isbn = {978-3-031-16868-0},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances},
pages = {237--255},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {In fact, common optimization problems in ENAS are computationally expensive and are usually handled using surrogate-assisted EAs(SAEAs) [1], employing inexpensive approximation regression and classification models, such as the Gaussian process model [2], radial basis network (RBN), etc., to replace the costly fitness evaluation [3]. SAEAs have proven to be useful and efficient in a variety of practical optimization applications [1].},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Sun, Yanan; Yen, Gary G.; Zhang, Mengjie
Hybrid GA and PSO for Architecture Design Book Chapter
In: Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances, pp. 171–180, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-16868-0.
@inbook{Sun2023b,
title = {Hybrid GA and PSO for Architecture Design},
author = {Yanan Sun and Gary G. Yen and Mengjie Zhang},
url = {https://doi.org/10.1007/978-3-031-16868-0_9},
doi = {10.1007/978-3-031-16868-0_9},
isbn = {978-3-031-16868-0},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances},
pages = {171--180},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {In this chapter, a new approach based on EC is introduced for automatically searching for the optimal CNN architecture and determining whether or not to use shortcut connections between one layer and its forward layer. After that, a two-level encoding strategy is applied to a hybrid EC methodology that is composed of a GA and a PSO. This allows for the generation of both the network architecture and the shortcut connections within it. The technique is referred to as DynamicNet because to the fact that during the course of the evolutionary process, both the architecture and the shortcut connections are determined dynamically without any involvement from a human being. On three widely used datasets that have differing degrees of complexity, DynamicNet will be evaluated in comparison with one method that is based on EC and 12 methods that are considered to be state-of-the-art.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Sun, Yanan; Yen, Gary G.; Zhang, Mengjie
Architecture Design for Skip-Connection Based CNNs Book Chapter
In: Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances, pp. 147–170, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-16868-0.
@inbook{Sun2023c,
title = {Architecture Design for Skip-Connection Based CNNs},
author = {Yanan Sun and Gary G. Yen and Mengjie Zhang},
url = {https://doi.org/10.1007/978-3-031-16868-0_8},
doi = {10.1007/978-3-031-16868-0_8},
isbn = {978-3-031-16868-0},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances},
pages = {147--170},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {In this chapter, an efficient and effective algorithms employing GA is introduced, dubbed CNN-GA, to find the best CNN architectures for specific image classification tasks automatically, such that the found CNN can be directly employed without any need for manual tuning. CNN-GA is an algorithm for automating the architecture design of CNN. Please keep in note that the terms ``automatic'' and ``automatic + manually + tuning'' are discussed from the perspective of end-users, rather than developers. In developing high-performance CNN architecture design algorithms, however, adequate domain expertise should be promoted. This effort is not difficult to comprehend by comparing it to the design of the Windows Operating System by Microsoft scientists: to ensure the users could be able to effectively operate on computers even if they do not have considerable understanding of operating systems, the scientists should put as much of their professional knowledge as they possibly can while building a user-friendly operating system.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Sun, Yanan; Yen, Gary G.; Zhang, Mengjie
Differential Evolution for Architecture Design Book Chapter
In: Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances, pp. 193–202, Springer International Publishing, Cham, 2023, ISBN: 978-3-031-16868-0.
@inbook{Sun2023d,
title = {Differential Evolution for Architecture Design},
author = {Yanan Sun and Gary G. Yen and Mengjie Zhang},
url = {https://doi.org/10.1007/978-3-031-16868-0_11},
doi = {10.1007/978-3-031-16868-0_11},
isbn = {978-3-031-16868-0},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Evolutionary Deep Neural Architecture Search: Fundamentals, Methods, and Recent Advances},
pages = {193--202},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {The general goal of this chapter is to explore the capacity of DE, named DECNN, to evolve deep CNN architectures and parameters automatically. Designing new crossover and mutation operators of DE, as well as an encoding scheme, and a second crossover operator will help to achieve the goal. DECNN will be evaluated on six datasets of various complexity that are widely used and compared to 12 state-of-the-art methods.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Jing, Kun; Chen, Luoyu; Xu, Jungang
An architecture entropy regularizer for differentiable neural architecture search Journal Article
In: Neural Networks, vol. 158, pp. 111-120, 2023, ISSN: 0893-6080.
@article{JING2023111,
title = {An architecture entropy regularizer for differentiable neural architecture search},
author = {Kun Jing and Luoyu Chen and Jungang Xu},
url = {https://www.sciencedirect.com/science/article/pii/S0893608022004567},
doi = {https://doi.org/10.1016/j.neunet.2022.11.015},
issn = {0893-6080},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Neural Networks},
volume = {158},
pages = {111-120},
abstract = {Differentiable architecture search (DARTS) is one of the prevailing paradigms of neural architecture search (NAS) due to allowing efficient gradient-based optimization during the search phase. However, its poor stability and generalizability are intolerable. We argue that the crux is the locally optimal architecture parameter caused by a dilemma, which is that the solutions to the Matthew effect and discretization discrepancy are inconsistent. To escape from the dilemma, we propose an architecture entropy to measure the discrepancy of the architecture parameters of different candidate operations and use it as a regularizer to control the learning of architecture parameters. Extensive experiments show that an architecture entropy regularizer with a negative or positive coefficient can effectively solve one side of the contradiction respectively, and the regularizer with a variable coefficient can relieve DARTS from the dilemma. Experimental results demonstrate that our architecture entropy regularizer can significantly improve different differentiable NAS algorithms on different datasets and different search spaces. Furthermore, we also achieve more accurate and more robust results on CIFAR-10 and ImageNet. The code is publicly available at https://github.com/kunjing96/DARTS-AER.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Yao, Fengqin; Wang, Shengke; Ding, Laihui; Zhong, Guoqiang; Bullock, Leon Bevan; Xu, Zhiwei; Dong, Junyu
Lightweight network learning with Zero-Shot Neural Architecture Search for UAV images Journal Article
In: Knowledge-Based Systems, vol. 260, pp. 110142, 2023, ISSN: 0950-7051.
@article{YAO2023110142,
title = {Lightweight network learning with Zero-Shot Neural Architecture Search for UAV images},
author = {Fengqin Yao and Shengke Wang and Laihui Ding and Guoqiang Zhong and Leon Bevan Bullock and Zhiwei Xu and Junyu Dong},
url = {https://www.sciencedirect.com/science/article/pii/S0950705122012382},
doi = {https://doi.org/10.1016/j.knosys.2022.110142},
issn = {0950-7051},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Knowledge-Based Systems},
volume = {260},
pages = {110142},
abstract = {Lightweight Network Architecture is essential for autonomous and intelligent monitoring of Unmanned Aerial Vehicles (UAVs), such as in object detection, image segmentation, and crowd counting applications. The state-of-the-art lightweight network learning based on Neural Architecture Search (NAS) usually costs enormous computation resources. Alternatively, low-performance embedded platforms and high-resolution drone images pose a challenge for lightweight network learning. To alleviate this problem, this paper proposes a new lightweight object detection model, called GhostShuffleNet (GSNet), for UAV images, which is built based on Zero-Shot Neural Architecture Search. This paper also introduces the new components which compose GSNet, namely GhostShuffle units (loosely based on ShuffleNetV2) and the backbone GSmodel-L. Firstly, a lightweight search space is constructed with the GhostShuffle (GS) units to reduce the parameters and floating-point operations (FLOPs). Secondly, the parameters, FLOPs, layers, and memory access cost (MAC) as constraints add to search strategy on a Zero-Shot Neural structure search algorithm, which then searches for an optimal network GSmodel-L. Finally, the optimal GSmodel-L is used as the backbone network and a Ghost-PAN feature fusion module and detection heads are added to complete the design of the lightweight object detection network (GSNet). Extensive experiments are conducted on the VisDrone2019 (14.92%mAP) dataset and the our UAV-OUC-DET (8.38%mAP) dataset demonstrating the efficiency and effectiveness of GSNet. The completed code is available at: https://github.com/yfq-yy/GSNet.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wang, Wenna; Zhang, Xiuwei; Cui, Hengfei; Yin, Hanlin; Zhang, Yannnig
FP-DARTS: Fast parallel differentiable neural architecture search for image classification Journal Article
In: Pattern Recognition, vol. 136, pp. 109193, 2023, ISSN: 0031-3203.
@article{WANG2023109193,
title = {FP-DARTS: Fast parallel differentiable neural architecture search for image classification},
author = {Wenna Wang and Xiuwei Zhang and Hengfei Cui and Hanlin Yin and Yannnig Zhang},
url = {https://www.sciencedirect.com/science/article/pii/S0031320322006720},
doi = {https://doi.org/10.1016/j.patcog.2022.109193},
issn = {0031-3203},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Pattern Recognition},
volume = {136},
pages = {109193},
abstract = {Neural Architecture Search (NAS) has made remarkable progress in automatic machine learning. However, it still suffers massive computing overheads limiting its wide applications. In this paper, we present an efficient search method, Fast Parallel Differential Neural Architecture Search (FP-DARTS). The proposed method is carefully designed from three levels to construct and train the super-network. Firstly, at the operation-level, to reduce the computational burden, different from the standard DARTS search space (8 operations), we decompose the operation set into two non-overlapping operator sub-sets (4 operations for each). Adopting these two reduced search spaces, two over-parameterized sub-networks are constructed. Secondly, at the channel-level, the partially-connected strategy is adopted, where each sub-network only adopts partial channels. Then these two sub-networks construct a two-parallel-path super-network by addition. Thirdly, at the training-level, the binary gate is introduced to control whether a path participates in the super-network training. It may suffer an unfair issue when using softmax to select the best input for intermediate nodes across two operator sub-sets. To tackle this problem, the sigmoid function is introduced, which measures the performance of operations without compression. Extensive experiments demonstrate the effectiveness of the proposed algorithm. Specifically, FP-DARTS achieves 2.50% test error with only 0.08 GPU-days on CIFAR10, and a state-of-the-art top-1 error rate of 23.7% on ImageNet using only 2.44 GPU-days for search.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Jin, Cong; Huang, Jinjie; Wei, Tianshu; Chen, Yuanjian
Neural architecture search based on dual attention mechanism for image classification Journal Article
In: Mathematical Biosciences and Engineering, vol. 20, no. 2, pp. 2691-2715, 2023, ISSN: 1551-0018.
@article{nokey,
title = {Neural architecture search based on dual attention mechanism for image classification},
author = {Cong Jin and Jinjie Huang and Tianshu Wei and Yuanjian Chen},
url = {https://www.aimspress.com/article/doi/10.3934/mbe.2023126},
doi = {10.3934/mbe.2023126},
issn = {1551-0018},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {Mathematical Biosciences and Engineering},
volume = {20},
number = {2},
pages = {2691-2715},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Jin, Yaochu; Zhu, Hangyu; Xu, Jinjin; Chen, Yang
Evolutionary Multi-objective Federated Learning Book Chapter
In: Federated Learning: Fundamentals and Advances, pp. 139–164, Springer Nature Singapore, Singapore, 2023, ISBN: 978-981-19-7083-2.
@inbook{Jin2023,
title = {Evolutionary Multi-objective Federated Learning},
author = {Yaochu Jin and Hangyu Zhu and Jinjin Xu and Yang Chen},
url = {https://link.springer.com/chapter/10.1007/978-981-19-7083-2_3},
doi = {10.1007/978-981-19-7083-2_3},
isbn = {978-981-19-7083-2},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Federated Learning: Fundamentals and Advances},
pages = {139--164},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Different from model quantization and partial model uploads presented in the previous chapter, evolutionary federated learning, more specifically, evolutionary federated neural architecture search, aims to optimize the architecture of neural network models, thereby reducing the communication costs caused by frequent model transmissions, generating lightweight neural models that are better suited for mobile and other edge devices, and also enhancing the final global model performance. To achieve this, scalable and efficient encoding methods for deep neural architectures must be designed and evolved using multi-objective evolutionary algorithms. This chapter presents two multi-objective evolutionary algorithms for federated neural architecture search. The first one employs a probabilistic representation of deep neural architectures that describes the connectivity between two neighboring layers and simultaneously maximizing the performance and minimizing the complexity of the neural architectures using a multi-objective evolutionary algorithm. However, this evolutionary framework is not practical for real-time optimization of the neural architectures in a federated environment. To tackle this challenge, a real-time federated evolutionary neural architecture search is then introduced. In addition to adopting a different neural search space, a double sampling strategy, including sampling subnetworks from a pretrained supernet and sampling clients for model update, is proposed so that the performance of the neural architectures becomes more stable, and each client needs to train one local model in one communication round, thereby preventing sudden performance drops during the optimization and avoiding training multiple submodels in one communication round. This way, evolutionary neural architecture search is made practical for real-time real-world applications.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}