Skip to content

tensorflow implementation of CondenseNet: An Efficient DenseNet using Learned Group Convolutions

Notifications You must be signed in to change notification settings

markdtw/condensenet-tensorflow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CondenseNet tensorflow

Tensorflow implementation of CondenseNet: An Efficient DenseNet using Learned Group Convolutions. The code is tested with cifar10, inference phase not implemented yet.

Model architecture

Official PyTorch implementation by @ShichenLiu here.

Prerequisites

Data

Preparation

  • Go to data/ folder and run python2 generate_cifar10_tfrecords.py --data-dir=./cifar-10-data. This code is directly borrowed from tensorflow official repo and have to be run with python 2.7+.

Train

Use default parameters:

python main.py

Check out tunable hyper-parameters:

python main.py --help

Other parameters including stages, groups, condense factor, and growth rate are in experiment.py.

Notes

  • Training for 300 epochs with the default settings reach testing accuracy 93.389% (paper report is 94.94%). There might be some details I didn't notice, feel free to point them out.
  • All the default parameters settings follow the paper/official pytorch implementation.
  • Current implmentations of standard group convolution and learned group convolution are very inefficient (a bunch of reshape, transpose and concat), looking for help to build much more efficient graph.
  • Evaluation phase (index select) has not been implemented yet, looking for potential help as well :D.
  • Issues are welcome!

Resources

About

tensorflow implementation of CondenseNet: An Efficient DenseNet using Learned Group Convolutions

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages