I'm real busy...
I have been developing some auto-differentiation stuff with rust-ad, when that has full support I plan to entirely rework Cogent to be more general purpose like PyTorch and also include all the functionality of simple_optimization to make a fully featured optimization library (you might consider ML a sub-category of optimization).
Cogent is a very basic library for training basic neural networks for classification tasks. It is designed to be as simple as feasible. Hyperparameters in neural network training can be set automatically, so why not? Ideally you could simply do:
let net = NeuralNetwork::Train(&data);
This is the most basic and not quite there yet implementation of that idea.
Training a network to classify MNIST:
// Uses
use cogent::{
NeuralNetwork,
EvaluationData,MeasuredCondition
};
use ndarray::{Array2,Axis};
// Setup
// ----------
// 784-ReLU->800-Softmax->10
let mut net = NeuralNetwork::new(784,&[
Layer::Dense(800,Activation::ReLU),
Layer::Dense(10,Activation::Softmax)
]);
// Setting training and testing data.
// `get_mnist_dataset(bool)` simply gets MNIST dataset.
// The boolean specifies if it is the MNIST testing data (`true`) or training data (`false`).
// Sets training and testing data.
let (mut train_data,mut train_labels):(Array2<f32>,Array2<usize>) = get_mnist_dataset(false);
let (test_data,test_labels):(Array2<f32>,Array2<usize>) = get_mnist_dataset(true);
// Execution
// ----------
// Trains until no notable accuracy improvements are being made over a number of iterations.
// By default this would end training if 0.5% accuracy improvement was not seen over 12 iterations/epochs.
net.train(&mut train_data,&mut train_labels)
.evaluation_data(EvaluationData::Actual(&test_data,&test_labels)) // Sets evaluation data
.l2(0.1) // Implements L2 regularisation with a 0.1 lambda vlaue
.tracking() // Prints backpropgation progress within each iteration
.log_interval(MeasuredCondition::Iteration(1)) // Prints evaluation after each iteration
.go();
// If evaluation data is not set manually it will simply shuffle and split off a random group from training data to be evaluation data.
// In the case of MNIST where training and evaluation datasets are given seperately, it makes sense to set it as such.
// Evaluation
// ----------
let (cost,correctly_classified):(f32,u32) = net.evaluate(&test_data,&test_labels,None); // (cost,examples correctly classified)
println!("Cost: {:.2}",cost);
println!(
"Accuracy: {}/{} ({:.2}%)",
correctly_classified,
test_data.len_of(Axis(1)),
correctly_classified as f32 / test_data.len_of(Axis(1)) as f32
);
While a huge amount of my work has gone into making this and learning the basics of neural networks along the way, I am immensely (and I cannot stress this enough) amateur in inumerable ways.
If you find any issues I would really appreciate if you could let me know (and possibly suggest any solutions).
- GPU compute using ArrayFire Rust Bindings
- Optimisers: Stochastic gradient descent.
- Layers: Dense, Dropout
- Activations: Softmax, Sigmoid, Tanh, ReLU.
- Loss functions: Mean sqaured error, Cross entropy.
- Misc: L2 regularisation
- Setup ArrayFire Rust bindings (Ignore step 4).
- Add
cogent = "^0.6"
toCargo.toml
.
TODO (See note)
- Convolutional layers.
- Meticulous testing (making sure things work).
- Optimise usage of VRAM.
- Meticulous benchmarking (making sure things are fast).
- Benchmarking against other popular neural network libraries (Keras etc.)
- Automatic net creation: Bayesian optimisation of everything.
Please note that things may not be developed inorder, it is only my estimation.