RATs-NAS: Redirection of Adjacent Trails on Graph Convolutional Networks for Predictor-Based Neural Architecture Search

Yu Ming Zhang, Jun Wei Hsieh, Chun Chieh Lee, Kuo Chin Fan

Research output: Contribution to journalArticlepeer-review

Abstract

Manually designed convolutional neural networks (CNNs) architectures such as visual geometry group network (VGG), ResNet, DenseNet, and MobileNet have achieved high performance across various tasks, but design them is time-consuming and costly. Neural architecture search (NAS) automates the discovery of effective CNN architectures, reducing the need for experts. However, evaluating candidate architectures requires significant graphics processing unit (GPU) resources, leading to the use of predictor-based NAS, such as graph convolutional networks (GCN), which is the popular option to construct predictors. However, we discover that, even though the ability of GCN mimics the propagation of features of real architectures, the binary nature of the adjacency matrix limits its effectiveness. To address this, we propose redirection of adjacent trails (RATs), which adaptively learns trail weights within the adjacency matrix. Our RATs-GCN outperform other predictors by dynamically adjusting trail weights after each graph convolution layer. Additionally, the proposed divide search sampling (DSS) strategy, based on the observation of cell-based NAS that architectures with similar floating point operations (FLOPs) perform similarly, enhances search efficiency. Our RATs-NAS, which combine RATs-GCN and DSS, shows significant improvements over other predictor-based NAS methods on NASBench-101, NASBench-201, and NASBench-301.

Original languageEnglish
Pages (from-to)6672-6682
Number of pages11
JournalIEEE Transactions on Artificial Intelligence
Volume5
Issue number12
DOIs
StatePublished - 2024

Keywords

  • Cell-based NAS
  • neural architecture search (NAS)
  • predictor-based NAS

Fingerprint

Dive into the research topics of 'RATs-NAS: Redirection of Adjacent Trails on Graph Convolutional Networks for Predictor-Based Neural Architecture Search'. Together they form a unique fingerprint.

Cite this