Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

Classifying images from Google Quick Draw

NotificationsYou must be signed in to change notification settings

kolszewska/QuickDrawClassifier

Repository files navigation

Quickdraw Preview

Classifying images from Google Quick Draw dataset using customConvolutional Neural Network.

What is Google Quick Draw?

About the Google Quick Draw:

Quick, Draw! is an online game developed by Google that challenges players todraw a picture of an object or idea and then uses a neural network artificial intelligenceto guess what the drawings represent. The AI learns from each drawing, increasing itsability to guess correctly in the future. The game is similar to Pictionary in thatthe player only has a limited time to draw (20 seconds). The concepts that it guessescan be simple, like 'foot', or more complicated, like 'animal migration'. This game isone of many simple games created by Google that are AI based as part of a project knownas 'A.I. Experiments'.

Open Soure dataset is availablehere.

How to prepare data for training ?

First make sure that you have the correct data. You can downloadit from the link above. I have used theNumpy bitmap files (.npy) format.

For my experiment I have chosen the following10 classes:

  • Bear,
  • Bee,
  • Camel,
  • Cat,
  • Cow,
  • Crab,
  • Crocodile,
  • Dog,
  • Dolphin,
  • Duck.

For our training we need the PNG files so we need to convert them from the numpy format. To do this,firstly create a virtualenv:

$ virtualenv venv

Enter created virtualenv:

$source venv/bin/activate

Install needed dependencies:

$ (venv) pip install -r requirements.txt

Now we can useconvert_to_png.py script. It takes the datafromin folder and converts the defined number of numpy filesto the PNG images.

For my experiment I have decided to have:

  • 20 000 images intraining set,
  • 4 000 images invalidation set,
  • 1 000 images intesting set.

To convert our data, simply run:

$ (venv) python convert_to_png.py

How to train

To start training the network simply run:

$ (venv) python training/convolutional.py<batch_size>

Viewing results

You can view the training and validation accuracies andlosses in the TensorBoard. During each training the datais saved to theruns directory.

During the training of after it, you can simply start up theTensorBoard with the following command:

$ (venv) tensorboard --logdir runs

Classification visualisation

To get this kind of visualisation, simply head to theClassificationVisualisation.ipynb

Visualisation

Conducted experiments

I have conducted three experiments to see how would the network train with different values of thelearning rate andbatch size.

Tested the following combinations:

  • Learning rate=0.01 witch batch sizes=32, 64, 128, 256, 512
  • Learning rate=0.001 witch batch sizes=32, 64, 128, 256, 512
  • Learning rate=0.0001 witch batch sizes=32, 64, 128, 256, 512

Shared parameters during the experiments:

  • Optimizer:Adam
  • Epochs:50
  • Dropout:0.5

First experiment (lr=0.01)

Ex1 train accuracyEx1 validation accuracy

Ex1 train lossEx1 validation loss

Test accuracy:

Batch sizeAccuracyColor on the plot
3210%Grey
6410%Blue
12810%Red
25610%Orange
51265.13%Green

Summary: In this experiment we can see that choosing such large large learning rate makes itunable to actually hit a global minima. We can then observe the outcome of this: every model below the 512batch size was returning the same class which (when we have 10 classes) results in exactly 10% accuracy.

Second experiment (lr=0.001)

Ex2 train accuracyEx2 validation accuracy

Ex2 train lossEx2 validation loss

Test accuracy:

Batch sizeAccuracyColor on the plot
3283.82%Orange
6484.91%Blue
12885.78%Red
25685.58%Light blue
51285.84%Magenta

Summary: In this experiment we can see that decreasing the learning rate 10x have a massive effect for the testaccuracy.

Third experiment (lr=0.0001)

Ex3 train accuracyEx3 validation accuracy

Ex3 train lossEx3 validation loss

Test accuracy:

Batch sizeAccuracyColor on the plot
3285.77%Magenta
6486.11%Grey
12885.54%Orange
25685.34%Green
51284.57%Light blue

Summary: In this experiment we can observe that decreasing learning rate even more did not improved the testaccuracy dramaticaly, but it did change the plots. They are much smoother due to fact that we are making much smallersteps than before when searching for a global minima.


[8]ページ先頭

©2009-2025 Movatter.jp