0% found this document useful (0 votes)
82 views

Smart Bin Using Machine Learning

The purpose of this paper is to develop a computer vision model using deep learning technique that can detect the different kind of waste. Waste disposal and its management are considered as an essential part in maintaining cleanliness in the cities. Waste management becomes easy if segregation of different kind of waste happens at initial level. The paper begins with analyzing advantages and disadvantages of existing smart bins that mainly focus on weight of the waste inside them.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
82 views

Smart Bin Using Machine Learning

The purpose of this paper is to develop a computer vision model using deep learning technique that can detect the different kind of waste. Waste disposal and its management are considered as an essential part in maintaining cleanliness in the cities. Waste management becomes easy if segregation of different kind of waste happens at initial level. The paper begins with analyzing advantages and disadvantages of existing smart bins that mainly focus on weight of the waste inside them.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

8 VII July 2020

https://doi.org/10.22214/ijraset.2020.30585
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

Smart Bin using Machine Learning


Mekhla Tiwary1, Ashish Kumar2, Kaushal Kumar3, Kavyashree Nayak K4, Rajeshwari J5
1, 2, 3, 4, 5
Information Science and Engineering Dayananda Sagar College of Engineering, Bangalore, India

Abstract: The purpose of this paper is to develop a computer vision model using deep learning technique that can detect the
different kind of waste. Waste disposal and its management are considered as an essential part in maintaining cleanliness in the
cities. Waste management becomes easy if segregation of different kind of waste happens at initial level. The paper begins with
analyzing advantages and disadvantages of existing smart bins that mainly focus on weight of the waste inside them. There are
few proposed framework for waste segregation but a better and a strong model can be developed using deep learning. The goal is
to come up with an optimized object detection architecture may be used for segregating three different kinds of waste that are
normally generated, that is plastic, paper and wet waste in real-time with better accuracy and lesser time taken for generating the
inference.
The superiority of the architecture will be demonstrated by developing a deep learning model that recognizes waste belonging to
different classes. The model will be conveyed to recognize waste utilizing web-cam in an effective way and apply object location
to each frame.
Keywords: CNN, Machine Learning
I. INTRODUCTION
Detachment and a short time later reusing of waste materials is essential for an acceptable society. The current confinement what's
more, reusing structures anticipates that workplaces should sort waste by hand and use a movement of huge channels to detach out
extra described items. The motivation is to find a modified strategy for orchestrating waste. This can make getting ready plants
logically beneficial and help decline waste, as it isn't commonly the circumstance that the delegates sort everything with 100%
precision. This won't simply have positive biological effects yet moreover helpful money related effects. The Municipal specialists
keep up dustbins at better places in the entire city. It is their commitment to check and clear the waste kept in the dustbins at
standard between times. In any case, various on different occasions they turn up late or go with hardly anything as there may not be
adequate waste in the dustbin. If they are late, there may be a couple of chances of the defilement of the waste. It would incite the
advancement of infinitesimal living beings and diseases. The collected rubbish would then make air sullying and cause respiratory
issues like COPD, asthma, etc.
II. METHODOLOGY
Dustbins are the holders that are generally utilized for gathering family unit squander all around the globe. In our everyday life, we
dispose assortment of waste materials classified as mechanical waste, sewage squander, household squander, and so on. Dustbins
are utilized for gathering the waste materials. Dustbins that are placed along roadside contain different kinds of waste. Some are
biodegradable and some are not. It takes a lot of time and man power if we want to segregate the waste after collection. The solution
for this problem is to initially segregate the waste while collecting. It is possible only if we can recognize the types of waste
initially. Our brain instantly recognizes the objects contained when we’re shown an image. A lot of time is taken and huge amount
of training data is required for a machine to identify these objects. However, with the ongoing advances in the field of deep learning
and hardware aspects, the computer vision field has become more intuitive. Caffe is a deep learning system that is made with
articulation, speed, as well as modularity. Expressive design energizes application and development. Models and advancement are
characterized by configuration with no hard-coding. Switch among CPU and a GPU by setting a solidary banner to prepare on the
GPU machine at that points end to product bunches or cell phones. Convolutional neural networks are a special type of feed-forward
networks. These models are designed to emulate the behavior of a visual cortex. CNNs perform very well on visual recognition
tasks. CNNs have special layers called convolutional layers and pooling layers that allow the network to encode certain images
properties. The simplest architecture of a convolutional neural networks starts with an input layer (images) followed by a sequence
of convolutional layers and pooling layers, and ends with fully-connected layers. The convolutional layers are usually followed by
one layer of ReLU activation functions. The convolutional, pooling and ReLU layers act as learnable features extractors, while the
fully connected layers acts as a machine learning classifier. Furthermore, the early layers of the network encode generic patterns of
the images, while later layers encode the details patterns of the images.

©IJRASET: All Rights are Reserved 1555


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

After training the dataset, we are using Caffe model for predicting the type of waste. Caffe model is a deep learning framework
developed by the Berkeley Vision and Learning Centre(BVLC). It is written in C++ and has Python and MATLAB bindings. There
are 4 steps in training a CNN using a Caffe model:
1) Data Preparation: This is the step, we clean the images and store them in a format that can be used by Caffe model. We will
then write a Python script that will handle both image pre-processing and storage.
2) Model Definition: In this step, we choose a CNN architecture and define its parameters in a configuration file with extension
‘.prototxt’.
3) Solver Definition: This is responsible for model optimization. We define the solver parameters in a configuration file with
extension ‘.prototxt’.
4) Model Training: We then train the model by executing one Caffe command from the terminal. After training the model, we will
get the trained model in a file with extension‘.caffemodel’.
III. CNN- CONVOLUTION NEURAL NETWORK
The Convolutional neural systems (CNN) is profound learning strategy, as of late it has ventured forward and clear improvement in
the field of PC vision, for example, a picture division, object discovery, acknowledgment, and subtitling. Undoubtedly, it naturally
motivated by the human cerebrum. Convolutional nets and other related models under the profound learning umbrella are, best case
scenario practically identical to the neural systems exist in the human cerebrum. Much the same as the structure and tasks of the
natural neurons in the human visual cortex through the sending of various leveled multi-layers arranges, the comparative way the
profound neural system must be uncover powerful to learn the different plans of highlight portrayal getting from preparing
information. Its tasks are programmed and includes building errands could be resolve with all the more quick and solid way. They
are very ready to discover and viably use explicit quirks of picture classes on the off chance that a monstrous preparing dataset is
given. In the mid-90s, CNN was first propelled for the reason perceiving written by hand digits. Later on, in 2012 a significant
advancement was made by discharging AlexNet. There is an essential rule that needs to be considered in a unique case for
multilayer perceptron where each neuron is associated with the open field situated in forward-face. Additionally, the neurons have a
place with each layer in the system has similar loads. For the errand of item acknowledgment, it might be isolated into two principle
parts: object acknowledgment and item location. In this paper, we just spotlight on multi-class object acknowledgment. Be that as it
may, stretching out existing acknowledgment models to multiclass objects identification task needs to adjust the design of the
model. Tensorflow library offers full-help for preparing, testing, tuning and encourage model’s organization with very much
recorded models for every one of these assignments. We effectively applied five layers CNN model for acknowledgment with non-
direct enactment work Rectified Linear Unit (ReLU) for acknowledgment purposes. We instated predispositions with 0 worth and
for the underlying size of the loads introduction Wij at each layer, the heuristic methodology recommended was embraced as
standard instatement to plan the proposed CNN model:
IV. NEURAL NETWORK
The promising scholarly life is to be estimated through learning capacity. ML is currently competent to learn and anticipate with a
progressively modern approach to arrange any obscure marvel from given datasets. Artificial neural network (ANNs) is neuro-
naturally motivated. The human cerebrum is made out of complex multi-layers nerve cells in type of neurons. The convolutedness
of genuine neurons is exceptionally disconnected, yet ANNs is represented through programming model dependent on some
observational information and in this way, it licenses the PCs to learn and gives some anticipated Numerous models are motivated
by ANNs are Convolutional Neural Networks, Recurrent Neural Networks, Profound Belief Networks. Artificial neural network
(ANNs) is neuro-organically motivated. The human mind is made out of complex multi-layers nerve cells in type of neurons. The
convolutedness of genuine neurons is profoundly preoccupied, yet ANNs is administered through programming original dependent
on some observational information.
V. SALIENT OBJECT DETECTION
Visual saliency identification, one of the most significant and testing errands in PC vision, expects to feature the most prevailing
article districts in a picture. Various applications join the visual saliency to improve their execution, Comprehensively, there are two
parts of approaches in striking object discovery, specifically base up (BU) and top-down (TD). Neighborhood include differentiate
assumes the focal job in BU striking article recognition, paying little heed to the semantic substance of the scene. To learn
neighborhood highlight, differentiate, different nearby and worldwide highlights are separated from pixels, for example edges [129],
spatial data [130]. Be that as it may, elevated level and multi-scale semantic data can't be investigated with these low-level
highlights.

©IJRASET: All Rights are Reserved 1556


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

Therefore, low difference striking maps rather than remarkable articles are acquired. TD notable item recognition is task oriented
and takes earlier information about article classes to control the age of notable maps. Taking semantic division for instance, a
saliency map is created in the division to allot pixels to specific item classes by means of a TD approach. In a word, TD saliency can
be seen as a focal point of-consideration component, which prunes BU striking focuses that are probably not going to be portions of
the article. Because of the centrality for giving significant level and multiscale highlight portrayal in many corresponded PC vision
undertakings, for example, semantic division, edge location and nonexclusive article recognition, it is attainable and important to
stretch out CNN to remarkable article discovery. The early work by Eleonora Vig et follows a totally programmed information
driven way to deal with play out a largescale scan for ideal highlights, to be specific a group of profound systems with various
layers and parameters. To address the issue of restricted preparing information, Kummerer et al. proposed the Profound Gaze by
moving from the AlexNet to create a high dimensional component space and make a saliency map. A comparative engineering was
proposed by Huang et al. to incorporate saliency forecast into pre-prepared article acknowledgment DNNs. The exchange is
cultivated by tweaking DNNs' loads with a target work dependent on the saliency assessment measurements, for example,
Similarity, KL-Divergence and Standardized Scanpath Saliency.

VI. PROMISING FUTURE DIRECTIONS AND TASKS


Regardless of fast turn of events and accomplished promising progress of item location, there are as yet many open issues for future
work. The first is little item recognition, for example, happening in COCO dataset and in face recognition task. To improve
limitation precision on little items under fractional impediments, it is important to alter organize designs from the following aspect.
Due to the relationships between various undertakings inside and outside item location, perform multiple tasks joint streamlining
has just been concentrated by numerous analysts. Notwithstanding, it is attractive to thoroughly consider the attributes of various
sub-assignments of item location (for example super pixel semantic division in remarkable article identification) and stretch out
perform various tasks enhancement to different applications, for example, occurrence division, multi-object following and multi-
individual posture estimation [S4]. In addition, given a particular application, the data from various modalities, such as content,
warm information and pictures, can be intertwined to accomplish a more discriminant organize.

VII. SYSTEM DESIGN


The meaning of the engineering of a framework, segments, modules, interfaces, and information for a framework to satisfy
determined prerequisites is system design. System design can likewise be viewed as the utilization of frameworks hypothesis to item
advancement. The plan stage creates the general structure of the product. The objective of configuration eliminate is to calculate the
modules that ought to be in the framework to satisfy all the framework necessities in a proficient way. It will contain the subtleties
of every one of these modules, their working with different modules and the ideal yield from every module. The yield of the system
design process is a depiction of the product design.

VIII. SYSTEM ARCHITECTURE DESIGN


The definition of the structure and operation and more views of a system is known as system architecture. A formal description and
rendition of a system, organized in a way so that it supports reasoning about the working and behaviors of the system is called
architecture description. System architecture contains framework parts that cooperate and actualize the general framework as shown
in Fig-1.

Fig-1 System Architecture

©IJRASET: All Rights are Reserved 1557


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

IX. XML DATASET


A markup language that characterizes a lot of rules for encoding records in a configuration that is both intelligible just as machine-
coherent is known as Extensible Markup Language. Accentuating ease, comprehensive explanation, and convenience over the
Internet are the structure targets of XML. It is a printed data bunch with strong assistance through the Unicode for different human
vernaculars. The language is comprehensively used for the depiction of abstract data structures, for instance, those used in web
organizations, anyway the arrangement of XML revolves around records. Few layout systems exist to help in the importance of
XML-based lingos, while the product engineers have made various application programming interfaces (APIs) to help the treatment
of XML dataset.XML stores data in plain substance design. This gives an item and hardware self-sufficient strategy for taking care
of, delivery, similarly as sharing the data. With XML, data can be open to a wide scope of scrutinizing machines like people, PCs,
voice machines, news channels and a couple others.XML dataset addresses extraction of information of enthusiasm through a
simple to utilize plaintext affirmation that follows the structure of the XML document. The declaration is indented that matches the
XML structure and the data we are interested in is tagged against a dataset.

X. DATA FLOW DIAGRAM


A data flow diagram is fundamentally a graphical portrayal of the progression of information through a data framework. DFD has
been demonstrated exceptionally valuable in understanding a framework and can be effectively utilized during examination. A DFD
shows the progression of information through the given framework. It sees a framework as a capacity that changes the contributions
to the ideal yields. Any perplexing frameworks won't play out this change in one stage. The information that goes about as info will
commonly experience a progression of changes before it turns into the output. With an information stream outline appeared in Fig-
2, customers can envision how the structure will function and what will it accomplish, how the system will be completed. Old
framework information stream outlines can be drawn up and diverged from another frameworks information stream chart to pull in
relationships with execute a dynamically powerful framework. Data stream outlines are generally used to outfit the end customer
with a physical idea of where they had given data, finally as an effect upon the structure of the whole framework. In the point of
view of advancement of Data Flow Diagram (DFD). it is a unique outline type which lets graphically represent the "stream" of
information through different application part. This is shown in Fig-3 with the help of a Level 1 DFD. Henceforth, the Data Flow
Diagrams can be effectively utilized for perception of information preparing or organized structure.

Fig-2 Level 0 DFD

Fig-3 Level 1 DFD

©IJRASET: All Rights are Reserved 1558


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

XI. USE CASE DIAGRAM


The external objects that interact directly with the system are called as actors. Actors include humans, external devices and the other
software systems. The important thing about actors is that they are not under control of the application. In case of this paper, user of
the system is the actor. To find use cases, for each actor, list the fundamentally different ways in which the actor uses the system.
Each of these ways is called a use case.
Fig-4 illustrates the use case diagram.

Fig-4 Use Case Diagram

XII. FLOW CHART


Fig-5 depicts a simple flow chart for this paper. Flowcharts are used in documenting and designing simple processes or programs.
Like other types of diagrams, they help visualize what is going on and thereby help understand any process, and perhaps also find
the flaws, bottlenecks, and other less-obvious features within it. There are many different types of flowcharts. Each type has its own
repertoire of boxes and notational conventions. The two most common types of boxes present in a flowchart are as follows:
A processing step, usually called activity and is denoted as a rectangular box. A decision step, which is usually denoted as a
diamond.

Fig-5 Flow Chart

XIII. SEQUENCE DIAGRAM


Fig-6 below shows a sequence diagram and depicts different associations. An arrangement outline in a Unified Modeling Language
(UML) is a kind of sequence diagram. It shows how strategies work with one another and in what demand.

Fig-6 Sequence Diagram

It shows the individuals in an association and the game plan of messages among them. Each part is consigned a fragment in a table.

©IJRASET: All Rights are Reserved 1559


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

XIV. IMPLEMENTATION
The realizing of an application and the execution of any plan, idea, model, design, specification, policy, algorithm or standard is
known as implementation. At the end of the day, an implementation is an acknowledgment of a technical specification or an
algorithm as a program, programming segment, or other PC framework through programming and through deployment. Numerous
executions may exist for a given determination or for a given norm. Implementation is one of the most important phases of the
Software Development Life Cycle (SDLC). It encompasses all the processes that are involved in getting new software or hardware
operating properly in its environment. This includes installation, configuration, running, testing and finally making necessary
changes. Specifically, it involves coding the system using a particular programming language and transferring the design into an
actual working system.
XV. CAFFE MODEL OVERVIEW
After training the dataset, we are using Caffe model for predicting the type of waste. Caffe model is a deep learning framework
developed by the Berkeley Vision and Learning Centre(BVLC). It is written in C++ and has Python and MATLAB bindings. There
are 4 steps in training a CNN using a Caffe model:
1) Data Preparation: This is the step, we clean the images and store them in a format that can be used by Caffe model. We will
then write a Python script that will handle both image pre-processing and storage.
2) Model Definition: In this step, we choose a CNN architecture and define its parameters in a configuration file with extension
‘.prototxt’.
3) Solver Definition: This is responsible for model optimization. We define the solver parameters in a configuration file with
extension ‘.prototxt’.
4) Model Training: We then train the model by executing one Caffe command from the terminal.

XVI. USABALITY ASPECT


Python is a deciphered, elevated level, broadly useful programming language. It was made by Guido van Rossum and was first
discharged in 1991. Python has an alternate structure reasoning that underlines code intelligibility, quite utilizing critical
whitespaces. It gives builds that empower clear programming on both little just as enormous scales. Python highlights a unique tape
framework and programmed memory the executives. It bolsters different programming ideal models, including object-arranged,
basic, utilitarian and procedural. It additionally has a huge and complete standard library. Python translators are accessible for some
working frameworks. CPython, which is the reference execution of Python, is open-source programming and furthermore has a
network-based advancement model. Both Python and CPython are overseen by the non-benefit Python Software Foundation.
Instead of having the entirety of its usefulness incorporated with its center, Python was intended to be exceptionally extensible. This
smaller seclusion has made it especially well known as a method for adding programmable interfaces to almost all the current
applications. Van Rossum's vision of a little center language with a huge standard library and effectively extensible mediator
originated from his disappointments with ABC, which embraced the contrary methodology.
XVII. TECHNICAL ASPECT
Anaconda is a free and an open-source distribution of Python and R programming languages for scientific computing (data science,
machine learning applications, large-scale data processing, predictive analytics, etc.), that aims at simplifying package management
and deployment. Package versions are always managed by the package management system conda. The Anaconda circulation is
utilized by more than 13 million clients and incorporates in excess of 1500 well known information science bundles appropriate for
Windows, Linux, and MacOS.
XVIII. RESULTS
The main python file having name‘real_time_object_detection.py’ is stored in the object detection folder that also contains .prototxt
and .caffemodel files. This is essentially the real model that has been utilized for the item identification. List of the strings namely
‘paper’, ‘plastic’ and ‘wet waste’ that are utilized to include correct labels for each box are provided. Processing for a single image
is done and a reframe is required to decipher veil from box coordinates to picture coordinates and fit the picture size.
The given model captures only those objects that have been trained. The captured object is then identified as ‘paper’, ‘plastic’ or
‘wet waste’ accordingly. Identification of the kind of waste thus helps in the initial segregation of waste. All outputs are float32
NumPy arrays and are converted to appropriate types. Handles to input and output tensors are obtained and dimensions are
expanded since the model expects images to have shape. Actual detection is done and the results are visualized. The figures from
Fig-7 to Fig-10 are the screenshots taken while running the model.

©IJRASET: All Rights are Reserved 1560


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

A. Results using Caffe Model

Fig-7 Plastic Detection Using Caffe Model

Fig-8 Paper Detection using Caffe Model

Fig-9 Paper and Wet Waste Detection in a single frame

©IJRASET: All Rights are Reserved 1561


International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.429
Volume 8 Issue VII July 2020- Available at www.ijraset.com

Fig-10 Two Plastic Objects Detection using Caffe Model

XIX. CONCLUSION
Various works have been continuing to diminish proportion of waste total and to keep up and mastermind the waste present in the
holder. As needs be, by completing these splendid holders all around the world, the canisters will be anything but difficult to utilize,
and there will be sterile condition around the container. It will similarly be useful for the experts who can enlighten the stressed to
shield the dustbin from getting flood consequently human checking is decreased. Using this, we can screen the absolute waste
evacuation in a capable way. An Infra-red Sensor structure is accessible in the canister to recognize objects put around the dustbin.
This structure will offer alert sound hints when we keep garbage's around the buildup holder. This along these lines will diminish
the time the dustbin is stuffed, and thusly will serve significant for the overall population and the earth and ecological components
where we live for the improvement of our future. We have adequately achieved the investigation go after insightful trashcan and
shut with some charming results. The level discoverer is giving a better than average estimation of heights and showing its status
with different concealing drove. Likewise, the region sensor present at the outside of trashcan is thoroughly prepared to perceive
near to objects and decisively opening its top for near to objects. The alert organization of GSM is successfully running in the whole
system and is getting promising admonition as showed by the status of trashcan. So, all the data is assembled and taken care of
cleverly in an adaptable application that we could also use for future redesigns in the structure. As demonstrated by the results got,
this all out structure is viably completed for a singular trashcan. The proposed structure could be proficient with facilitated
arrangement of various trashcans each having its own GSM.

REFERENCES
[1] P. F. Felzenszwalb, R. B. Girshick, D. Mcallester, and D. Ramanan, “Object detection with discriminatively trained part-based models,” IEEE Trans. Pattern
Anal. Mach. Intell., vol. 32, no. 9, p. 1627, 2018.
[2] K. K. Sung and T. Poggio, “Example-based learning for view-based human face detection,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 20, no. 1, pp. 39–51,
2019.
[3] C. Wojek, P. Dollar, B. Schiele, and P. Perona, “Pedestrian detection: An evaluation of the state of the art,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no.
4, p. 743, 2017.
[4] H. Kobatake and Y. Yoshinaga, “Detection of spicules on mammogram based on skeleton analysis.” IEEE Trans. Med. Imag., vol. 15, no. 3, pp. 235–245,
2016.
[5] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, “Caffe: Convolutional architecture for fast feature
embedding,” in ACM MM, 2018.
[6] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in NIPS, 2016.
[7] Stuhlsatz, J. Lippel, and T. Zielke, “Feature extraction with deep neural networks by a generalized discriminant analysis.” IEEE Trans. Neural Netw. &
Learning Syst., vol. 23, no. 4, pp. 596–608, 2015.
[8] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in CVPR, 2014.
[9] L. Deng, M. L. Seltzer, D. Yu, A. Acero, A.-r. Mohamed, and G. Hinton, “Binary coding of speech spectrograms using a deep autoencoder,” in
INTERSPEECH, 215
[10] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv:1409.1556, 2016.
[11] W. Ouyang, X. Wang, X. Zeng, S. Qiu, P. Luo, Y. Tian, H. Li, S. Yang, Z. Wang, C.-C. Loy et al., “Deepid-net: Deformable deep convolutional neural
networks for object detection,” in CVPR, 2017.
[12] J. Xue, J. Li, and Y. Gong, “Restructuring of deep neural network acoustic models with singular value decomposition.” in Interspeech, 2018
[13] A. Pentina, V. Sharmanska, and C. H. Lampert, “Curriculum learning of multiple tasks,” in CVPR,2019

©IJRASET: All Rights are Reserved 1562

You might also like