-
Notifications
You must be signed in to change notification settings - Fork 19.6k
a capsule cnn on cifar-10 #9193
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 4 commits
81531af
ab0b308
6c53e54
9ae1789
8ef2397
f9cc297
6594763
4f1cbb9
307de84
25ff097
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,211 @@ | ||
'''Train a simple CNN-Capsule Network on the CIFAR10 small images dataset. | ||
paper: https://arxiv.org/abs/1710.09829 | ||
|
||
Without Data Augmentation: | ||
It gets to 75% validation accuracy in 10 epochs, | ||
and 79% after 15 epochs, and overfitting after 20 epcohs | ||
|
||
With Data Augmentation: | ||
It gets to 75% validation accuracy in 10 epochs, | ||
and 79% after 15 epochs, and 83% after 30 epcohs. | ||
In my test, highest validation accuracy is 83.79% after 50 epcohs. | ||
|
||
This is a fast Implement, just 20s/epcoh with a gtx 1070 gpu. | ||
''' | ||
|
||
|
||
# the Capsule Implement is from https://github.com/bojone/Capsule/ | ||
|
||
from __future__ import print_function | ||
from keras import backend as K | ||
from keras.engine.topology import Layer | ||
from keras.layers import Activation | ||
from keras import utils | ||
from keras.datasets import cifar10 | ||
from keras.models import Model | ||
from keras.layers import * | ||
from keras.preprocessing.image import ImageDataGenerator | ||
|
||
|
||
# a squashing function. but it has litte difference from the Hinton's paper. | ||
# it seems that this form of squashing performs better. | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Explain what the difference is and what motivates it |
||
def squash(x, axis=-1): | ||
s_squared_norm = K.sum(K.square(x), axis, keepdims=True) + K.epsilon() | ||
scale = K.sqrt(s_squared_norm) / (0.5 + s_squared_norm) | ||
return scale * x | ||
|
||
|
||
# define our own softmax function instead of K.softmax | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Why though? |
||
def softmax(x, axis=-1): | ||
ex = K.exp(x - K.max(x, axis=axis, keepdims=True)) | ||
return ex / K.sum(ex, axis=axis, keepdims=True) | ||
|
||
|
||
''' | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Please put this comment on top of your file with the description of the example. |
||
A Capsule Implement with Pure Keras | ||
There are two vesions of Capsule. | ||
One is like dense layer (for the fixed-shape input), | ||
and the other one is like Time distributed dense (for various length input). | ||
The input shape of Capsule must be (batch_size, | ||
input_num_capsule, | ||
input_dim_capsule | ||
) | ||
and the output shape is (batch_size, | ||
num_capsule, | ||
dim_capsule | ||
) | ||
''' | ||
|
||
|
||
class Capsule(Layer): | ||
def __init__(self, | ||
num_capsule, # the number of output capsules | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Please put the documentation in the docstring and follow Keras' documentation style. You can find the style in the docstring of any layer. |
||
dim_capsule, # the dim of output capsules | ||
routings=3, # the iter of dynamic routing | ||
share_weights=True, # share_weights or not | ||
activation='default', # it can use our own activation | ||
# rather than squashing. | ||
# 'default' is squashing. | ||
**kwargs): | ||
super(Capsule, self).__init__(**kwargs) | ||
self.num_capsule = num_capsule | ||
self.dim_capsule = dim_capsule | ||
self.routings = routings | ||
self.share_weights = share_weights | ||
if activation == 'default': | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Why default? What does that mean? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. default means we use squashing function as activation. I rename it now. |
||
self.activation = squash | ||
else: | ||
# the activation is compatible with keras. | ||
# e.g. we can set activation='relu' | ||
self.activation = Activation(activation) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This should be |
||
|
||
def build(self, input_shape): | ||
super(Capsule, self).build(input_shape) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. No point in calling super here |
||
input_dim_capsule = input_shape[-1] | ||
if self.share_weights: | ||
self.W = self.add_weight(name='capsule_kernel', | ||
shape=(1, input_dim_capsule, | ||
self.num_capsule * | ||
self.dim_capsule), | ||
initializer='glorot_uniform', | ||
trainable=True) | ||
else: | ||
input_num_capsule = input_shape[-2] | ||
self.W = self.add_weight(name='capsule_kernel', | ||
shape=(input_num_capsule, | ||
input_dim_capsule, | ||
self.num_capsule * | ||
self.dim_capsule), | ||
initializer='glorot_uniform', | ||
trainable=True) | ||
|
||
def call(self, u_vecs): | ||
# it is very important to use K.conv1d or K.local_conv1d | ||
# to get the fast speech. NOT use K.map_fn! | ||
if self.share_weights: | ||
u_hat_vecs = K.conv1d(u_vecs, self.W) | ||
else: | ||
u_hat_vecs = K.local_conv1d(u_vecs, self.W, [1], [1]) | ||
|
||
batch_size = K.shape(u_vecs)[0] | ||
input_num_capsule = K.shape(u_vecs)[1] | ||
u_hat_vecs = K.reshape(u_hat_vecs, (batch_size, | ||
input_num_capsule, | ||
self.num_capsule, | ||
self.dim_capsule)) | ||
u_hat_vecs = K.permute_dimensions(u_hat_vecs, (0, 2, 1, 3)) | ||
# final u_hat_vecs.shape = [None, num_capsule, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Are those comments needed? |
||
# input_num_capsule, dim_capsule] | ||
|
||
b = K.zeros_like(u_hat_vecs[:, :, :, 0]) # shape = [None, num_capsule, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Put your comment on top of the operation please. |
||
# input_num_capsule] | ||
for i in range(self.routings): | ||
c = softmax(b, 1) | ||
outputs = self.activation(K.batch_dot(c, u_hat_vecs, [2, 2])) | ||
if i < self.routings - 1: | ||
b = K.batch_dot(outputs, u_hat_vecs, [2, 3]) | ||
|
||
return outputs | ||
|
||
def compute_output_shape(self, input_shape): | ||
return (None, self.num_capsule, self.dim_capsule) | ||
|
||
|
||
# some parameters | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Clarify or remove. |
||
batch_size = 128 | ||
num_classes = 10 | ||
epochs = 100 | ||
(x_train, y_train), (x_test, y_test) = cifar10.load_data() | ||
|
||
x_train = x_train.astype('float32') | ||
x_test = x_test.astype('float32') | ||
x_train /= 255 | ||
x_test /= 255 | ||
y_train = utils.to_categorical(y_train, num_classes) | ||
y_test = utils.to_categorical(y_test, num_classes) | ||
|
||
# A common Conv2D model | ||
input_image = Input(shape=(None, None, 3)) | ||
cnn = Conv2D(64, (3, 3), activation='relu')(input_image) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
|
||
cnn = Conv2D(64, (3, 3), activation='relu')(cnn) | ||
cnn = AveragePooling2D((2, 2))(cnn) | ||
cnn = Conv2D(128, (3, 3), activation='relu')(cnn) | ||
cnn = Conv2D(128, (3, 3), activation='relu')(cnn) | ||
|
||
''' | ||
now we reshape it as (batch_size, input_num_capsule, input_dim_capsule) | ||
then connect a Capsule layer. | ||
the output of final model is the lengths of 10 Capsule, who dim=16 | ||
the length of Capsule is the proba, | ||
so the probelm becomes a 10 two-classification problems | ||
''' | ||
|
||
cnn = Reshape((-1, 128))(cnn) | ||
capsule = Capsule(10, 16, 3, True)(cnn) | ||
output = Lambda(lambda x: K.sqrt(K.sum(K.square(x), 2)))(capsule) | ||
model = Model(inputs=input_image, outputs=output) | ||
|
||
# we use a margin loss | ||
model.compile(loss=lambda y_true, y_pred: y_true * K.relu(0.9 - y_pred)**2 + | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. It is a good idea to implement your loss outside of compile so that your code is readable. |
||
0.25 * (1 - y_true) * K.relu(y_pred - 0.1)**2, | ||
optimizer='adam', | ||
metrics=['accuracy']) | ||
model.summary() | ||
|
||
# we can compare the perfermace with or without data augmentation | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Typo: perfermace -> performance |
||
data_augmentation = True | ||
|
||
if not data_augmentation: | ||
print('Not using data augmentation.') | ||
model.fit(x_train, y_train, | ||
batch_size=batch_size, | ||
epochs=epochs, | ||
validation_data=(x_test, y_test), | ||
shuffle=True) | ||
else: | ||
print('Using real-time data augmentation.') | ||
# This will do preprocessing and realtime data augmentation: | ||
datagen = ImageDataGenerator( | ||
featurewise_center=False, # set input mean to 0 over the dataset | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Everything in the There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This is copied verbatim from another example. If comments are removed, it'll be inconsistent with other examples, where this is pervasive throughout. |
||
samplewise_center=False, # set each sample mean to 0 | ||
featurewise_std_normalization=False, # divide inputs by dataset std | ||
samplewise_std_normalization=False, # divide each input by its std | ||
zca_whitening=False, # apply ZCA whitening | ||
rotation_range=0, # randomly rotate images in 0 to 180 degrees | ||
width_shift_range=0.1, # randomly shift images horizontally | ||
# (fraction of total width) | ||
height_shift_range=0.1, # randomly shift images vertically | ||
# (fraction of total height) | ||
horizontal_flip=True, # randomly flip images | ||
vertical_flip=False) # randomly flip images | ||
|
||
# Compute quantities required for feature-wise normalization | ||
# (std, mean, and principal components if ZCA whitening is applied). | ||
datagen.fit(x_train) | ||
|
||
# Fit the model on the batches generated by datagen.flow(). | ||
model.fit_generator(datagen.flow(x_train, y_train, | ||
batch_size=batch_size), | ||
epochs=epochs, | ||
validation_data=(x_test, y_test), | ||
workers=4) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
epochs
(typo)