3.2 Preprocessing
3.2 Preprocessing
2 PREPROCESSING:
Here we have used auto resizing for training to make all the images in the
dataset to convert in to same resolution.
The process of feature extraction is useful when you need to reduce the
number of resources needed for processing without losing important or
relevant information. Feature extraction can also reduce the amount of
redundant data for a given analysis. Also, the reduction of the data and
the machine’s efforts in building variable combinations (features)
facilitate the speed of learning and generalization steps in the machine
learning process.
3.5 CNN:
The CNN can deal with those input only for what the neural network is
trained and the data is saved.
DATASET:
One major advantage of using CNNs over NNs is that you do not need to flatten
the input images to 1D as they are capable of working with image data in 2D.
This helps in retaining the “spatial” properties of images.So here we are using
x-ray data base which consist of three categories Frontal,Maxillary and Normal
PRE-PROCESSING STEPS
Activation function
TENSORFLOW IMPLEMENTATION:
The main components in a TensorFlow system are the client, which uses the
Session interface to communicate with the master, and one or more worker
processes, with each worker process responsible for arbitrating access to one
or more computational devices (such as CPU cores or GPU cards) and for
executing graph nodes on those devices as instructed by the master. We have
both local and distributed implementations of the TensorFlow interface. The
local implementation is used when the client, the master, and the worker all
run on a single machine in the context of a single operating system process
(possibly with multiple devices, if for example, the machine has many GPU
cards installed). The distributed implementation shares most of the code with
the local implementation, but extends it with support for an environment
where the client, the master, and the workers can all be in different processes
on different machines. In our distributed environment, these different tasks
are containers in jobs managed by a cluster scheduling system . These two
different modes are illustrated . Most of the rest of this section discusses issues
that are common to both implementations, while This discusses some issues
that are particular to the distributed implementation.
The VGG-16 and VGG-19 architecture consist of large kernel-size filters with
multiple 3×3 kernel-size filters, one after another. Within a given receptive field
(the effective area size of an input image on which output depends), multiple
stacked smaller sized kernels are better than a single larger sized kernel because
multiple non-linear layers increase the depth of the network, enabling it to learn
more complex features at a lower cost. As a result, the 3×3 kernels in the VGG
architecture help to retain more fine details of an image . The ResNet
architecture is similar to the VGG model, consisting mostly of 3×3 filters.
Additionally, the ResNet model has a network depth of as large as 152.
Therefore, it achieves better accuracy than VGG and GoogleNet, while being
computationally more efficient than VGG. While the VGG and ResNet models
achieve phenomenal accuracy, their deployment on even the most modest sized
GPUs is a problem because of the massive computational requirements, both in
terms of memory and time. There are several limitations to this study. First, the
external test dataset in multiple medical centers did not be included for
reproducibility. In the case of X-ray equipment, there is a relatively small
difference in performance compared to other medical imaging equipment,
depending on the manufacturer or model. In this study, therefore, the external
test dataset in other medical centers did not be included. In the case of a local
medical center using relatively old equipment; however, an additional
performance evaluation is also required to utilize artificial intelligence (AI)
assistive software. Second, the proposed majority decision algorithm was
optimized to evaluate only maxillary sinusitis. Therefore, there is a limitation to
evaluate sinusitis in frontal, ethmoid, and sphenoid. In order to utilize AI based
assistive software in the future, further study is underway because it is
necessary to evaluate sinusitis at other locations as well as maxillary. Third, it
lacks pattern recognition and representation methods that can solve black-box in
deep learning. It needs to determine a reasonable consensus for solving the
black-box problem. The feature recognition based activation map was used to
solve the black-box problem in deep learning. As it can be shown from the
results, not only classification but also lesion localization can be expressed as a
result. It helps medical doctors make a reasonable inference about the deep
learning analysis. However, it is not enough to understand all deep leaning
procedures. For example, it is difficult to understand the pattern of each learned
CNN model. By understanding the pattern recognition capabilities of each
model, we can understand the advantages and disadvantages of each model and
achieve the optimization of the overall AI system. To overcome this limitation,
a feature connectivity representation should be available for each layer to
determine which feature weights are strong. In addition to feature
representation, text-based description algorithm can be applied to overcome the
black-box limitation in a medical application using the convolutional recurrent
neural network (CRNN) that is the combination CNN and recurrent neural
network (RNN) (20,21). A majority decision algorithm with multiple CNN
models was shown to have high accuracy and significantly more accurate lesion
detection ability compared to individual CNN models. The proposed deep
learning method using PNS X-ray images can be used as an adjunct tool to help
improve the diagnostic accuracy of maxillary sinusitis.