Model: "model_9"
__________________________________________________________________________________________________
Layer (type) Output Shape Param # Connected to
==================================================================================================
img (InputLayer) (None, 128, 128, 1) 0
__________________________________________________________________________________________________
conv2d_207 (Conv2D) (None, 128, 128, 16) 160 img[0][0]
__________________________________________________________________________________________________
batch_normalization_196 (BatchN (None, 128, 128, 16) 64 conv2d_207[0][0]
__________________________________________________________________________________________________
activation_196 (Activation) (None, 128, 128, 16) 0 batch_normalization_196[0][0]
__________________________________________________________________________________________________
max_pooling2d_55 (MaxPooling2D) (None, 64, 64, 16) 0 activation_196[0][0]
__________________________________________________________________________________________________
dropout_86 (Dropout) (None, 64, 64, 16) 0 max_pooling2d_55[0][0]
__________________________________________________________________________________________________
conv2d_209 (Conv2D) (None, 64, 64, 32) 4640 dropout_86[0][0]
__________________________________________________________________________________________________
batch_normalization_198 (BatchN (None, 64, 64, 32) 128 conv2d_209[0][0]
__________________________________________________________________________________________________
activation_198 (Activation) (None, 64, 64, 32) 0 batch_normalization_198[0][0]
__________________________________________________________________________________________________
max_pooling2d_56 (MaxPooling2D) (None, 32, 32, 32) 0 activation_198[0][0]
__________________________________________________________________________________________________
dropout_87 (Dropout) (None, 32, 32, 32) 0 max_pooling2d_56[0][0]
__________________________________________________________________________________________________
conv2d_211 (Conv2D) (None, 32, 32, 64) 18496 dropout_87[0][0]
__________________________________________________________________________________________________
batch_normalization_200 (BatchN (None, 32, 32, 64) 256 conv2d_211[0][0]
__________________________________________________________________________________________________
activation_200 (Activation) (None, 32, 32, 64) 0 batch_normalization_200[0][0]
__________________________________________________________________________________________________
max_pooling2d_57 (MaxPooling2D) (None, 16, 16, 64) 0 activation_200[0][0]
__________________________________________________________________________________________________
dropout_88 (Dropout) (None, 16, 16, 64) 0 max_pooling2d_57[0][0]
__________________________________________________________________________________________________
conv2d_213 (Conv2D) (None, 16, 16, 128) 73856 dropout_88[0][0]
__________________________________________________________________________________________________
batch_normalization_202 (BatchN (None, 16, 16, 128) 512 conv2d_213[0][0]
__________________________________________________________________________________________________
activation_202 (Activation) (None, 16, 16, 128) 0 batch_normalization_202[0][0]
__________________________________________________________________________________________________
max_pooling2d_58 (MaxPooling2D) (None, 8, 8, 128) 0 activation_202[0][0]
__________________________________________________________________________________________________
dropout_89 (Dropout) (None, 8, 8, 128) 0 max_pooling2d_58[0][0]
__________________________________________________________________________________________________
conv2d_215 (Conv2D) (None, 8, 8, 256) 295168 dropout_89[0][0]
__________________________________________________________________________________________________
batch_normalization_204 (BatchN (None, 8, 8, 256) 1024 conv2d_215[0][0]
__________________________________________________________________________________________________
activation_204 (Activation) (None, 8, 8, 256) 0 batch_normalization_204[0][0]
__________________________________________________________________________________________________
conv2d_transpose_38 (Conv2DTran (None, 16, 16, 128) 295040 activation_204[0][0]
__________________________________________________________________________________________________
concatenate_38 (Concatenate) (None, 16, 16, 256) 0 conv2d_transpose_38[0][0]
activation_202[0][0]
__________________________________________________________________________________________________
dropout_90 (Dropout) (None, 16, 16, 256) 0 concatenate_38[0][0]
__________________________________________________________________________________________________
conv2d_217 (Conv2D) (None, 16, 16, 128) 295040 dropout_90[0][0]
__________________________________________________________________________________________________
batch_normalization_206 (BatchN (None, 16, 16, 128) 512 conv2d_217[0][0]
__________________________________________________________________________________________________
activation_206 (Activation) (None, 16, 16, 128) 0 batch_normalization_206[0][0]
__________________________________________________________________________________________________
conv2d_transpose_39 (Conv2DTran (None, 32, 32, 64) 73792 activation_206[0][0]
__________________________________________________________________________________________________
concatenate_39 (Concatenate) (None, 32, 32, 128) 0 conv2d_transpose_39[0][0]
activation_200[0][0]
__________________________________________________________________________________________________
dropout_91 (Dropout) (None, 32, 32, 128) 0 concatenate_39[0][0]
__________________________________________________________________________________________________
conv2d_219 (Conv2D) (None, 32, 32, 64) 73792 dropout_91[0][0]
__________________________________________________________________________________________________
batch_normalization_208 (BatchN (None, 32, 32, 64) 256 conv2d_219[0][0]
__________________________________________________________________________________________________
activation_208 (Activation) (None, 32, 32, 64) 0 batch_normalization_208[0][0]
__________________________________________________________________________________________________
conv2d_transpose_40 (Conv2DTran (None, 64, 64, 32) 18464 activation_208[0][0]
__________________________________________________________________________________________________
concatenate_40 (Concatenate) (None, 64, 64, 64) 0 conv2d_transpose_40[0][0]
activation_198[0][0]
__________________________________________________________________________________________________
dropout_92 (Dropout) (None, 64, 64, 64) 0 concatenate_40[0][0]
__________________________________________________________________________________________________
conv2d_221 (Conv2D) (None, 64, 64, 32) 18464 dropout_92[0][0]
__________________________________________________________________________________________________
batch_normalization_210 (BatchN (None, 64, 64, 32) 128 conv2d_221[0][0]
__________________________________________________________________________________________________
activation_210 (Activation) (None, 64, 64, 32) 0 batch_normalization_210[0][0]
__________________________________________________________________________________________________
conv2d_transpose_41 (Conv2DTran (None, 128, 128, 16) 4624 activation_210[0][0]
__________________________________________________________________________________________________
concatenate_41 (Concatenate) (None, 128, 128, 32) 0 conv2d_transpose_41[0][0]
activation_196[0][0]
__________________________________________________________________________________________________
dropout_93 (Dropout) (None, 128, 128, 32) 0 concatenate_41[0][0]
__________________________________________________________________________________________________
conv2d_223 (Conv2D) (None, 128, 128, 16) 4624 dropout_93[0][0]
__________________________________________________________________________________________________
batch_normalization_212 (BatchN (None, 128, 128, 16) 64 conv2d_223[0][0]
__________________________________________________________________________________________________
activation_212 (Activation) (None, 128, 128, 16) 0 batch_normalization_212[0][0]
__________________________________________________________________________________________________
conv2d_224 (Conv2D) (None, 128, 128, 1) 17 activation_212[0][0]
==================================================================================================
Total params: 1,179,121
Trainable params: 1,177,649
Non-trainable params: 1,472
__________________________________________________________________________________________________