def load_images(imagepath):
= []
image_data = []
image_data_flatten = []
labels = os.listdir(imagepath)
files_or_folders for i in files_or_folders:
if(os.path.isdir(imagepath+i)):
= os.listdir(imagepath+i)
imagefiles for j in imagefiles:
if (np.asarray(Image.open(imagepath+i+"/"+j)).shape == (224,224,3)):
= np.asarray(Image.open(imagepath+i+"/"+j))
data
image_data.append(data)= data.reshape(-1,)
data
image_data_flatten.append(data)
labels.append(i)return image_data, image_data_flatten, labels
Introduction
This article explores the use of convolution neural networks (CNN). Also, since it is my second deep learning project, I took every opportuinity to build my basics strongly.
Dataset consists of 15 vegetable ( Bean, Bitter_Gourd, Bottle_Gourd, Brinjal, Broccoli, Cabbage, Capsicum, Carrot, Cauliflower, Cucumber, Papaya, Potato, Pumpkin, Radish, Tomato) images resized to 224X224 pixel and arranged in three different folders of train, test and dev.
Each class has equal proportion of images making it a balanced dataset.
Traditional Machine Learning
First of all, I have used the de-facto image processing package of python - PIL
( Python Imaging Library ). Further, created an iterative function load_images
to load images from the respective directory.
Later, have modeled both Logistic Regression & Random Forest. Random Forest gave a good lift on test accuracy, but it is over-fitting on training data.
Model | Train Accuracy | Test/Validation Accuracy |
---|---|---|
Logistic Regression | 77% | 62% |
Random Forest | 100% | 79% |
Below is detailed classification matrix for both these models. ( Carrot identification better by both models )
Deep Learning/ Neural Networks
If I had performed my experimentation on same notebook and in one sitting, I would have used the loaded images output which was used earlier. However, as the learning continued for multiple days, I came across image_dataset_from_directory
which is far more easier to load the images in keras
.
Below is the simple command where we can load the images directly.
= tf.keras.utils.image_dataset_from_directory(
train_ds
train_dir,=123,
seed=(224, 224),
image_size=32)
batch_size
= tf.keras.utils.image_dataset_from_directory(
test_ds
test_dir,=123,
seed=(224, 224),
image_size=32)
batch_size
= tf.keras.utils.image_dataset_from_directory(
val_ds
val_dir,=123,
seed=(224, 224),
image_size=32) batch_size
One & Two Layered CONV2D Model
To start with I have explored Single
and Two
layered CONV2D Models.
= tf.keras.Sequential([
model 32, (3,3), activation='relu', input_shape=(224, 224, 3)),
tf.keras.layers.Conv2D(2,2),
tf.keras.layers.MaxPooling2D(
tf.keras.layers.Flatten(),512, activation='relu'),
tf.keras.layers.Dense(15, activation='softmax')
tf.keras.layers.Dense( ])
= tf.keras.Sequential([
model_layered 32, (5,5), activation='relu', input_shape=(224, 224, 3)),
tf.keras.layers.Conv2D(2,2),
tf.keras.layers.MaxPooling2D(64, (3,3), activation='relu'),
tf.keras.layers.Conv2D(2,2),
tf.keras.layers.MaxPooling2D(
tf.keras.layers.Flatten(),1024, activation='relu'),
tf.keras.layers.Dense(15, activation='softmax')
tf.keras.layers.Dense( ])
Model | Epochs | Train Accuracy | Test/Validation Accuracy |
---|---|---|---|
One-Layered | 10 | 97% | 63% |
Two-Layered | 10 | 90% | 52% |
Both these models, definetly over fits on the training data suggesting us to explore better models.
Rescaling
Rescaling helps to create a better model.
= tf.keras.Sequential([
model_rescale 1./255, input_shape=(224, 224, 3)),
tfl.Rescaling(32,7,padding='same',activation='relu'),
tfl.Conv2D(
tfl.MaxPooling2D(),64,5,padding='same',activation='relu'),
tfl.Conv2D(
tfl.MaxPooling2D(),128,3,padding='same',activation='relu'),
tfl.Conv2D(
tfl.MaxPooling2D(),
tfl.Flatten(),1024,activation='relu'),
tfl.Dense(128,activation='relu'),
tfl.Dense(15,activation='softmax')
tfl.Dense( ])
Model | Epochs | Train Accuracy | Test/Validation Accuracy |
---|---|---|---|
Re-scaling | 10 | 98% | 93% |
We can see that by reshaping the data and adding a layered CONV2D structure we got a way better model compared to traditional ML methods. However, still it over fits on the training data.
Data-Augmentation
The most common way to avoid overfitting is to perform either flipping, rotating or zoomning the image. The commands are fairly simple if we use keras
.
input = tf.keras.Input(shape=(224, 224, 3))
= tfl.RandomFlip("horizontal")(input)
x = tfl.RandomRotation(0.2)(x)
x = tfl.RandomZoom(0.2)(x)
x = tfl.Rescaling(1./255)(x)
x = tfl.Conv2D(32,7,padding='same',activation='relu')(x)
x = tfl.MaxPooling2D()(x)
x = tfl.Conv2D(64,7,padding='same',activation='relu')(x)
x = tfl.MaxPooling2D()(x)
x = tfl.Conv2D(128,7,padding='same',activation='relu')(x)
x = tfl.MaxPooling2D()(x)
x = tfl.Flatten()(x)
x = tfl.Dense(1024,activation='relu')(x)
x = tfl.Dense(512,activation='relu')(x)
x = tfl.Dense(15,activation = "softmax")(x)
output
= tf.keras.Model(input,output)
model_dataaug model_dataaug.summary()
Model | Epochs | Train Accuracy | Test/Validation Accuracy |
---|---|---|---|
Data Augmentation | 10 | 95% | 95% |
Wow!!! we got a very decent and balanced model with 95% accuracy with data augmentation. Now, Let’s explore transfer learning.
Transfer Learning with Custom Layers
Earlier, transfer learning of MobileNetV2
without custom layer has got an accuracy of 97% on test and dev sets. So, I have added two custom layers of Dense
with 1024 and 512 neurons each.
Here is the network structure of the transfer learning model.
= tf.keras.Input(shape=(224, 224, 3))
inputs = tfl.Rescaling(1./255)(inputs)
x = tfl.RandomFlip("horizontal")(x)
x = tfl.RandomRotation(0.2)(x)
x = tfl.RandomZoom(0.2)(x)
x = mobilenet_model(x, training=False)
x = tf.keras.layers.Dropout(0.2)(x)
x = tfl.Flatten()(x)
x = tfl.Dense(1024, activation='relu')(x)
x = tfl.Dropout(0.2)(x)
x = tfl.Dense(512, activation='relu')(x)
x = tfl.Dropout(0.2)(x)
x = tf.keras.layers.Dense(15)(x)
outputs = tf.keras.Model(inputs, outputs) model
And WOW!!! This model gave 100 % Accuracy. Wohooo!!!!
Model | Epochs | Train Accuracy | Test/Validation Accuracy |
---|---|---|---|
Transfer_Learning | 15 | 100% | 100% |
Below is detailed classification matrix for this model on test data.
Conclusion
It is wonderful to learn that CNN
models are far better for images. Adding to it, transfer learning with custom layers is always better to achieve higher accuracy in quick time. I am so happy to get 100% accuracy on my first exploration on image dataset.