What is Neural Architecture Search (NAS)?
The Problem It Solves
Manual architecture design is slow and requires deep expertise. An expert might try 50-100 architecture variations over months, but the search space contains millions of possible designs. Human intuition cannot exhaustively explore this space.
NAS automates this exploration. It evaluates thousands of architectures systematically, finding designs that humans might never consider. The result is often architectures that outperform hand-designed ones by 1-5% accuracy at the same or lower compute cost.
Three Core Components
Search space: Defines what architectures are possible. This includes layer types (convolution, attention, pooling), connections (skip connections, dense blocks), and sizes (channel counts, depths). Larger spaces find better architectures but cost more to search.
Search strategy: How to explore the space. Options include reinforcement learning, evolutionary algorithms, or gradient-based methods. Each has different exploration-exploitation tradeoffs.
Evaluation strategy: How to score each candidate architecture. Full training is expensive (hours to days per architecture). Multi-fidelity methods train partially or use proxies to estimate final accuracy cheaply.