NAS-Bench-101 Towards Reproducible Neural Architecture Search

Posted on May 29, 2022   1 minute read ∼ Filed in  : 

Introduction

The paper makes the following contributions:

  1. The paper introduces the NAS-Bench-101, which includes 423k unique architectures and their corresponding accuracy and runtime (Trained on cifar 10)

  2. Conducts experiments on various optimization algorithms on NAS.

    NASBench Dataset

    Architecture

    image-20220530143537272

    Cell Encoding

    Each cell is encoded by an adjacency matrix and a label vector with length 3, each element in the label vector is one operation. So there are a total of 510M unique cells => 510M unique models.

    Each cell in a model is the same.

    After removing the invalid module (no path from input or edges exceeds 9 ), there are 423K models left.

    Combine semantics

    When multiple edges point to the same vertex, the incoming tensors must be combined. Adding them or concatenating them are both standard techniques (NASI uses the weighted combination. )

    The paper uses the rule: tensors going to the output vertex are concatenated and those going into other vertices are summed.

    Training

    Datasets: CIFAR10, 40k training examples, 10k validation examples, 10k testing examples.

    Hyperparameters:

    1. utilize a single, fixed set of hyperparameters for all NAS-Bench-101 models. (50 in this paper. )
    2. Find hyperparameters optimizing the average accuracy of those models.

    Optimization: RMSProp

    Loss: cross-entry loss with L2 decay.

    Matrix

    The paper evaluates each architecture after training three times with different random initializations.

    1. Training accuracy;
    2. Validation accuracy;
    3. Testing accuracy;
    4. Training time in seconds;
    5. A number of trainable model parameters.

    NASBench as a Benchmark

    NAS algorithm compare

    The paper firstly compares various NAS algorithms and HPO algorithms including

    1. RS, SMAC, BOHB.
    2. regularized evaluation
    3. RL based

    Each algorithm runs 500 independent trails, (evaluate + update algorithm), and measure the regret vs time spent on evaluating those 500 trails.

image-20220530172250301

Foundings:

  1. RE, BOHB, and SMAC perform best and start to outperform RS.
  2. RE, BOHB, and SMAC show the most robust performance.




END OF POST




Tags Cloud


Categories Cloud




It's the niceties that make the difference fate gives us the hand, and we play the cards.