TS Nas
TS Nas
Bachelor of Technology
in
Computer Science and Engineering (AI&ML)
Submitted by
D. SIMHADRI
(21H51A66E6)
2024-2025
.
CMR COLLEGE OF ENGINEERING & TECHNOLOGY
KANDLAKOYA, MEDCHAL ROAD, HYDERABAD – 501401
CERTIFICATE
This is to certify that the Technical Seminar report entitled " Neural Architeture
Search(NAS)" being submitted by D. Simhadri (21H51A66E6) in partial
fulfillment for the award of Bachelor of Technology in Computer Science and
Engineering (AI&ML) is a record of bonafide work carried out his/her under my
guidance and supervision.
The results embodies in this project report have not been submitted to
any other University or Institute for the award of any Degree.
With great pleasure we want to take this opportunity to express my heartfelt gratitude to all
the people who helped in making this project work a grand success.
We are grateful to Ms. Sana Afreen, Assistant Professor, Department of Computer
Science and Engineering (AI&ML) for her valuable technical suggestions and guidance during the
execution of this project work.
We would like to thank Dr. P. Sruthi, Head of the Department of Computer Science and
Engineering (AI&ML), CMR College of Engineering and Technology, who is the major driving
forces to complete my project work successfully.
We are very grateful to Dr. Ghanta Devadasu, Dean-Academics, CMR College of
Engineering and Technology, for his constant support and motivation in carrying out the project
work successfully.
We extend our heartfelt gratitude to Dr. Seshu Kumar Avadhanam, Principal, CMR
College of Engineering & Technology, for his unwavering support and guidance in the successful
completion of our project and his encouragement has been invaluable throughout this endeavor.
We are highly indebted to Dr. V A Narayana, Director, CMR College of Engineering and
Technology, for giving permission to carry out this project in a successful and fruitful way.
We express our sincere thanks to Shri. Ch. Gopal Reddy, Secretary& Correspondent,
CMR Group of Institutions, and Shri Ch Abhinav Reddy, CEO, CMR Group of Institutions for
their continuous care and support.
We would like to thank the Teaching & Non- teaching staff of Department of Computer
Science and Engineering for their co-operation.
Finally, We extend thanks to our parents who stood behind us at different stages of this
Project. We sincerely acknowledge and thank all those who gave support directly and indirectly in
completion of this project work.
D. Simhadri 21H51A66E6
Neural Architecture Search (NAS)
TABLE OF CONTENTS
List of Figures
FIGURE
NO. TITLE PAGE NO.
List of Acronyms
ACRONYM
NO. ACRONYM FULL FORM
1 NAS Neural Architecture Search
2 GA Genetic Algorithm
3 AI Artificial Intelligence
4 EA Evolutionary Algorithm
5 RL Reinforcement Algorithm
6 DARTS Differentiable Architecture Search
7 AR Augmented Reality
ABSTRACT
Neural Architecture Search (NAS) is a method used to automate the design of deep neural
networks, reducing human intervention in the model selection process. By employing
techniques such as reinforcement learning, evolutionary algorithms, and gradient-based
methods, NAS optimizes network structures for specific tasks. This process enhances
efficiency, accelerates model development, and often results in architectures that outperform
manually designed networks. Despite its potential, NAS faces challenges such as high
computational costs and the complexity of search spaces. Recent advancements focus on
reducing these constraints through weight-sharing mechanisms, proxy tasks, and hardware-
aware searches. The future of NAS promises more adaptable and scalable architectures that
cater to diverse real-world applications. As NAS continues to evolve, it is expected to play a
crucial role in advancing deep learning by making neural network design more efficient,
scalable, and adaptable to various real-world applications.
Deep learning has led to significant advancements in areas like image recognition, speech
processing, and machine translation. A key factor behind this progress is the design of
innovative neural network architectures. Traditionally, these architectures have been crafted
manually by domain experts, which is both time-intensive and prone to human error. To
address this, the field is increasingly shifting toward automated methods for neural
architecture search (NAS). These approaches are typically categorized based on three
components: the search space, the search strategy, and the performance estimation strategy.
CHAPTER 1
INTRODUCTION
CHAPTER 1
INTRODUCTION
1.1 Introduction
In recent years, the field of deep learning has witnessed remarkable progress, thanks in part to
the development of advanced neural network architectures. Neural Architecture Search (NAS)
has emerged as a powerful technique to automate the process of designing effective neural
networks. In this tutorial, we’ll look into the fundamentals of NAS, exploring its concepts,
methodologies, and practical applications.
The success of deep learning in perceptual tasks is largely due to its automation of the feature
engineering process: hierarchical feature extractors are learned in an end-to-end fashion from
data rather than manually designed. This success has been accompanied, however, by a rising
demand for architecture engineering, where increasingly more complex neural architectures
are designed manually. Neural Architecture Search (NAS), the process of automating
architecture engineering, is thus a logical next step in automating machine learning. NAS can
be seen as subfield of AutoML and has significant overlap with hyperparameter optimization
and meta-learning We categorize methods for NAS according to three dimensions:
Key Contributions:
High Accuracy: AmoebaNet has achieved state-of-the-art performance on image
classification tasks, rivaling architectures designed through manual effort.
Evolutionary Process: It leverages evolutionary algorithms to explore the vast
architecture space without requiring hand-engineering.
3. Gradient-Based NAS
Gradient-based NAS uses continuous relaxation of the architecture search space to make it
differentiable, allowing the architecture to be optimized using gradient descent. One popular
example is DARTS (Differentiable Architecture Search).
Example: DARTS (Differentiable Architecture Search)
DARTS formulates the NAS problem as a continuous optimization task by relaxing discrete
architectural choices into weighted sums. This allows the architecture search to be optimized
using standard gradient-based optimization techniques, making it more efficient than RL or
EA approaches.
Key Contributions:
Efficiency: DARTS is more efficient than traditional NAS approaches, as it does not
require training a separate model for each architecture sampled.
Scalability: DARTS can scale to larger tasks due to its efficient optimization process.
CHAPTER 2
LITERATURE
REVIEW
CHAPTER 2
LITERATURE REVIEW
2.1 Literature Review
Hyperparameter optimization is an important research topic in machine learning, and is widely
used in practice (Bergstra et al., 2011; Bergstra & Bengio, 2012; Snoek et al., 2012; 2015;
Saxena & Verbeek, 2016). Despite their success, these methods are still limited in that they
only search models from a fixed-length space. In other words, it is difficult to ask them to
generate a variable-length configuration that specifies the structure and connectivity of a
network. In practice, these methods often work better if they are supplied with a good initial
model (Bergstra & Bengio, 2012; Snoek et al., 2012; 2015). There are Bayesian optimization
methods that allow to search non fixed length architectures (Bergstra et al., 2013; Mendoza et
al., 2016), but they are less general and less flexible than the method proposed in this paper.
Modern neuro-evolution algorithms, e.g., Wierstra et al. (2005); Floreano et al. (2008);
Stanley et al. (2009), on the other hand, are much more flexible for composing novel models,
yet they are usually less practical at a large scale. Their limitations lie in the fact that they are
search-based methods, thus they are slow or require many heuristics to work well.
Neural Architecture Search has some parallels to program synthesis and inductive
programming, the idea of searching a program from examples (Summers, 1977; Biermann,
1978). In machine learning, probabilistic program induction has been used successfully in
many settings, such as learning to solve simple Q&A (Liang et al., 2010; Neelakantan et al.,
2015; Andreas et al., 2016), sort a list of numbers (Reed & de Freitas, 2015), and learning with
very few examples (Lake et al., 2015).
Also related to our work is the idea of learning to learn or meta-learning (Thrun & Pratt,
2012), a general framework of using information learned in one task to improve a future task.
More closely related is the idea of using a neural network to learn the gradient descent updates
for another network (Andrychowicz et al., 2016).
The early years of Neural Architecture Search (2016-2017) were dominated by reinforcement
learning (RL)-based approaches, with the seminal work by Zoph & Le (2016) establishing the
foundational framework. Their method employed an RNN controller that generated neural
network architectures trained through policy gradients, achieving state-of-the-art performance
on CIFAR-10. However, this breakthrough came at an extraordinary computational cost,
requiring over 2,000 GPU days to complete the architecture search. Several follow-up works
attempted to refine this approach, including MetaQNN (Baker et al., 2016) which utilized Q-
learning for architecture search, and BlockQNN (Zhong et al., 2017) that implemented a
hierarchical search strategy with RL. While these methods demonstrated promising results,
they shared common limitations of prohibitive computational requirements and limited
scalability to larger datasets or more complex architectures, hindering their practical adoption.
The field saw significant advancement with the introduction of evolutionary algorithms (2017-
2018), which offered an alternative approach to architecture search. AmoebaNet (Real et al.,
2018) became a notable example, employing tournament selection to evolve neural
architectures that matched manually designed networks while using three times fewer
parameters. Another important contribution was the hierarchical evolutionary approach (Liu
et al., 2017), which simultaneously optimized both macro and micro-structures of neural
networks. These evolutionary methods provided better interpretability compared to RL-based
approaches, as the architecture development process followed more transparent selection and
mutation operations. However, they remained computationally intensive, though somewhat
less so than their RL predecessors, and still posed challenges for widespread implementation.
CHAPTER 3
PROPOSED
WORK
CHAPTER 3
PROPOSED WORK
3.1 Problem Definition
Designing effective neural network architectures is a critical but complex task in deep
learning. Traditionally, this process has required extensive human expertise, trial-and-error
experimentation, and domain-specific knowledge, making it both time-consuming and
resource-intensive. As deep learning models become increasingly complex and are applied to
a wider range of domains, the manual design of neural architectures becomes a significant
bottleneck in the development pipeline.
Neural Architecture Search (NAS) aims to automate the design of neural networks by
leveraging machine learning algorithms to discover optimal architectures for a given task.
However, existing NAS techniques often face several challenges:
1. High Computational Cost: Many NAS methods require thousands of GPU hours,
making them inaccessible for most researchers and impractical for rapid deployment.
2. Inefficient Search Methods: Search algorithms like reinforcement learning and
evolutionary strategies can be slow to converge or may explore suboptimal parts of the
search space.
3. Generalization Issues: Architectures discovered on smaller datasets (e.g., CIFAR-10)
may not perform well when transferred to larger, real-world datasets (e.g., ImageNet).
4. Complex Search Spaces: The design of search spaces significantly influences NAS
outcomes, yet defining a balanced and effective space remains a challenge.
5. Lack of Interpretability: NAS-generated models often act as black boxes, making it
hard to understand their design rationale or ensure trustworthiness in critical
applications.
3.2 Real Time Applications
Neural Architecture Search (NAS) is transforming how neural networks are designed,
enabling automatic discovery of optimized models for real-world tasks. In real-time
applications, where latency, speed, and adaptability are critical, NAS offers a powerful way to
create models that are not only accurate but also efficient in execution.
CMR College of Engineering and Technology CSE (AI&ML) 9
Neural Architecture Search (NAS)
1. Real-Time Image and Video Processing
Autonomous Vehicles
Use Case: Object detection, lane detection, pedestrian recognition, and traffic sign
recognition.
Role of NAS: Finds efficient CNN architectures for real-time video feeds with low latency
on edge devices.
Example: Lightweight NAS-designed models deployed on NVIDIA Jetson platforms for
real-time image inference.
Surveillance Systems
Use Case: Real-time human detection, activity recognition, and anomaly detection.
Role of NAS: Helps design models with fast inference speeds suitable for 24/7 operation
on low-power hardware.
Example: NAS-based object detection models in smart CCTV systems.
2. Mobile and Edge Computing
Smartphones and Wearables
Use Case: Face unlock, voice commands, camera enhancements, health tracking.
Role of NAS: Optimizes model size and speed for limited hardware (battery and processor
constraints).
Example: Google’s MNASNet uses NAS to balance accuracy and latency on mobile
devices for tasks like photo enhancement and AI assistants.
IoT Devices
Use Case: Real-time analytics on sensors, smart home automation, industrial IoT.
Role of NAS: Enables efficient models that run directly on the device without needing
cloud inference.
3. Natural Language Processing (NLP)
Voice Assistants (e.g., Alexa, Siri, Google Assistant)
Use Case: Real-time speech recognition, intent detection, and response generation.
Role of NAS: Speeds up transformer model search to fit memory and latency budgets.
Customer Support Chatbots
Use Case: Real-time question answering, summarization, and feedback analysis.
Role of NAS: Automates finding compact transformer variants for fast response
generation.
CMR College of Engineering and Technology CSE (AI&ML) 10
Neural Architecture Search (NAS)
You can refer to this excellent review by Esken et.al for more details. Also feel free to
advice the great article by Lillian Weng as an extra resource.
However, recent approaches combine the search strategy with the evaluation step, making it
hard to distinguish algorithms between them.
For that reason, we will explore NAS based solely on the search strategy. As we progress,
we will examine different search spaces and evaluation techniques. Also note that many
implementations experiment with different types or search strategies so the following
categorization is not always strict.
Search strategy refers to the methodology used to search for the optimal architecture in the
search space. We can classify NAS algorithms by their search strategy into 5 main areas:
1. Random search
2. Reinforcement learning
3. Evolutionary algorithms
4. Sequential model-based optimization
5. Gradient optimization
Random search
The most naive approach is obviously random search, which is often used as a baseline.
Here a valid architecture is chosen at random with no learning involved whatsoever.
Reinforcement learning
NAS can be very elegantly formulated as an RL problem. The agent’s action is the
generation of a neural architecture while the agent’s reward is the performance evaluation.
The action space is of course the search space. As a result, different RL methods can be used
to solve the problem.
Early works of NAS (NAS-RL, NASNet) used a recurrent neural network (RNN) as a policy
network (controller). The RNN is responsible for generating candidate architectures. The
architecture is then trained and evaluated on the validation set. The parameters of the RNN
controller are optimized in order to maximize the expected validation accuracy. How? Using
policy gradients techniques such as REINFORCE and Proximal Policy Optimization (PPO).
Similarly, ENAS uses an RNN controller trained with policy gradients. Notably, it is one of
the first works that effectively share parameters among architectures. The intuition is that the
CMR College of Engineering and Technology CSE (AI&ML) 13
Neural Architecture Search (NAS)
architectures can be viewed as part of a large graph, an approach that has been used
extensively as we will see below. ENAS training is performed in two alternating steps: a) the
RNN controller is trained with REINFORCE and b) the shared parameters are trained in
typical gradient descent form
Modular search space
To solve the global space problem, cell-based approaches were proposed in order to
“modularize” the search space. That is, mixing different blocks of layers called
modules. NASNet is the most popular algorithm in that category. NASNet learns only two
kinds of modules or “cells” : a normal cell that performs feature extraction and a
reduction cell that downsamples the input. The final architecture is built by stacking these
cells in a predefined way.
Evolutionary algorithms
Genetic Algorithms (GA) is an alternative way to optimize the network architecture.
Evolutionary algorithms start with a population of models. In each step, some models are
sampled and “reproduce” to generate offsprings by applying mutations to them. Mutations
can be local operations such as the addition of a layer, the modification of a hyperparameter,
etc. After training, they evaluated and added back to the population. The process repeats
itself until a certain condition is met.
3.4 Architecture
Fig:3.2 Architecture of NAS
CMR College of Engineering and Technology CSE (AI&ML) 14
CHAPTER 4
CONCLUSION
CMR College of Engineering and Technology CSE (AI&ML) 16
Neural Architecture Search (NAS)
CHAPTER 4
CONCLUSION
4.1 Conclusion and Future Enhancements
Neural Architecture Search (NAS) represents a significant advancement in the field of
automated machine learning, enabling the design of optimal neural network architectures
without extensive manual intervention. By leveraging techniques such as reinforcement
learning, evolutionary algorithms, and gradient-based methods, NAS can efficiently explore
vast search spaces to identify models that achieve superior performance on a variety of tasks.
Despite challenges like high computational cost and complexity, NAS has proven its potential
in generating state-of-the-art models for image classification, object detection, and natural
language processing. Moreover, hardware-aware and multi-objective NAS approaches have
further extended its applicability to real-time and resource-constrained environments.
As research in NAS continues to evolve, future improvements in efficiency, transferability,
and interpretability are likely to make this approach more accessible and practical. Overall,
NAS is paving the way toward fully automated AI model development, reducing human effort
while maintaining or even surpassing expert-designed architectures.
Neural Architecture Search (NAS) is transforming the field of AI by automating the design of
efficient and high-performing neural networks. From reinforcement learning-based approaches
like NASNet to gradient-based methods like DARTS, NAS is making it easier and faster to
discover innovative architectures for a wide range of tasks. While challenges like
computational cost and transferability remain, ongoing research and advancements like weight
sharing and one-shot NAS are making NAS more practical and scalable. As NAS continues to
evolve, it will play a crucial role in the future of AI development, enabling more efficient,
powerful, and accessible machine learning models.
REFERENCES
REFERENCES
[1]. Abdelfattah M.S, Mehrotra A, Dudziak Ł, Lane N.D. (2021). Zero-cost proxies for
lightweight NAS. arXiv preprint arXiv:2101.08134
[2]. Baker B, Gupta O, Raskar R, Naik N: (2017). Accelerating neural architecture search
using performance prediction. arXiv preprint arXiv:1705.10823
[3].Banbury C, Zhou C, Fedorov I, Matas R, Thakker U, Gope D, Janapa Reddi V, Mattina M,
Whatmough P (2021) Micronets: Neural network architectures for deploying tinyml
applications on commodity microcontrollers. Proc Machine Learn Syst 3:517–532
[4]. Barnell M, Raymond C, Smiley S, Isereau D, Brown D. (2022). Ultra low-power deep
learning applications at the edge with Jetson Orin AGX hardware. In: 2022 IEEE High
Performance Extreme Computing Conference (HPEC), pp. 1–4 IEEE
[5].Bender G, Kindermans P.-J, Zoph B, Vasudevan V, Le Q. (2018). Understanding and
simplifying one-shot architecture search. In: International Conference on Machine Learning,
pp. 550–559. PMLR
[6].Benmeziane H, Maghraoui KE, Ouarnoughi H, Niar S, Wistuba M, Wang N (2021) A
comprehensive survey on hardware-aware neural architecture search. arXiv
preprint arXiv:2101.09336
[7]. Caglar B, Broggi G, Ali MA, Orgéas L, Michaud V (2022) Deep learning accelerated
prediction of the permeability of fibrous microstructures. Composites Part Appl Sci Manufact
158:106973