Projectppt Edited
Projectppt Edited
U-NET
by
Keerthi Vundavalli
108121138
https://www.kaggle.com/datasets/nikhilroxtomar/brain-tumor-segmentation
Model architecture
U-Net, an image is converted into a vector and then the same mapping is used to convert it again to
an image. This reduce Using a combination of convolution, up-convolution, pooling and skip
connections, the U-Net is able to extract and capture complex features, while also keeping and
reconstructing spatial information the distortion by preserving the original structure of the image.
• Encoder: The network begins with a series of convolutional layers that downsample the input
image, extracting features and capturing context. This part is similar to the encoder in a typical
convolutional neural network (CNN).
• Bottleneck: At the bottom of the U-shape, the bottleneck consists of two convolutional layers
that further process the features extracted by the encoder. This stage serves as a bridge
between the encoder and decoder, retaining the most critical information while discarding the
less important details.
• Decoder: The decoder mirrors the encoder but uses upsampling layers instead of pooling layers.
Each upsampling step is followed by a concatenation with the corresponding feature map from the
encoder. This step ensures that fine-grained information lost during downsampling is recovered.
Convolutional layers are applied after concatenation to refine the segmentation map.
• Skip Connections: One of the key features of U-Net is the use of skip connections between
corresponding layers in the encoder and decoder. These connections concatenate features from
the encoder with those in the decoder, helping to preserve spatial details and improve
segmentation accuracy.
• Output Layer: The final layer of the U-Net is a 1x1 convolution that reduces the number of output
channels to the number of classes, typically followed by a softmax or sigmoid activation function
for multi-class or binary segmentation, respectively.
Block diagram of the UNET
architecture
Training
Data splitting divides a dataset into three main subsets:Training Set: The model is trained on this
data, learning patterns to make predictions.Validation Set: During training, the model is evaluated
on this data to fine-tune parameters and avoid overfitting, ensuring it generalizes well to new
data.Testing Set: After training and validation, the model is tested on this data to evaluate its
performance, providing an estimate of how it will perform in real-world scenarios.
• Training split:
Training data : 60%
Validation data : 20%
Test data : 20%
• Hyperparameters:
batch size = 5
learning rate = 1e-4
number of epochs = 10
• Framework: TensorFlow and Keras.
Results
• Evaluation metrics
The model was evaluated using the test set, and the following average scores were obtained:
F1: 0.941484716
Jaccard: 0.889438944
Recall: 0.990808824
Precision: 0.896838602
Visual results
• https://ieeexplore.ieee.org/document/9827132
• https://www.sciencedirect.com/science/article/pii/S1746809422008813