Skip to content

flexible and extensible implementation of a multithreaded feedforward neural network in Java including popular optimizers, wrapped up in a console user interface

Notifications You must be signed in to change notification settings

spaceshark123/NeuralNetwork

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

36 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Neural Network in Java

This project provides a flexible and extensible implementation of a multithreaded feedforward neural network in Java with popular optimizers included, wrapped up in a console user interface with a realtime accuracy graph visualizer while training. The neural network is designed to be easy to use and customize, making it a valuable tool for various machine learning and deep learning tasks. This was made without any external machine learning, math, or other libraries to aid in its creation (pure Java, at least for the base class).

Table of Contents

Overview

Neural networks are a fundamental building block of modern machine learning and artificial intelligence. This Java-based implementation allows you to create, train, evaluate, and use neural networks for a wide range of applications using a highly user-friendly console/GUI interface or directly using the NeuralNetwork class for your own needs if necessary.

Key Features

  • User-friendly Console Interface: Everything is wrapped in an easy-to-use console interface using commands to perform tasks, similar to a shell terminal. This allows for no-code experimentation with neural networks.

  • Realtime Accuracy Graph While Training: Displays realtime graph of accuracy over epochs for mnist networks while training, allowing users to visualize improvements as they occur.

  • Realtime MNIST Drawing Tool: Includes a drawing tool that allows users to draw on a 28x28 canvas and recieve realtime predictions by the network of which MNIST digit it is. This allows for easy debugging and evaluation.

  • Parallelism and Multithreading: Uses parallel computing and multithreading to dramatically accelerate computation time for evaluation and training mini-batches.

  • Training Callback Interface: Provides an interface that includes a callback method that can be passed into the Train function to execute custom code on each mini-batch iteration of training.

  • Customizable Topology: Define the number of neurons in each layer and activation functions for each layer.

  • Multiple Activation Functions: Supports popular activation functions including linear, sigmoid, tanh, relu, binary, and softmax.

  • Multiple Loss Functions: Supports major loss/error functions: Mean Squared Error, Sum Squared Error, and Categorical Cross-Entropy.

  • Multiple Optimizers: Includes popular optimizers like Stochastic gradient descent (SGD), SGD with momentum, AdaGrad, RMSProp, and Adam. Supports additional, custom optimizers using the NeuralNetwork.Optimizer interface.

  • Weight Initialization: Utilizes weight initialization techniques like Xavier (for linear, sigmoid, and tanh) and He (for relu) for better convergence.

  • Data Augmentation: Includes option to augment MNIST dataset on import, which performs random affine transformations (translation, rotation, scaling) to increase generalizability of networks.

  • Training: Train your neural network using mini-batched Gradient Descent (SGD) with customizable optimizer, loss function, learning rate, mini-batch size, and learning rate decay.

  • Train/Test Set Cross-Validation: During training, train and test set accuracy are measured for each mini-batch, ensuring a complete picture of network performance and prevention of overfitting. The console interface allows users to specify a train/test split ratio.

  • Gradient Clipping: Helps prevent exploding gradients during training by setting a gradient clipping threshold.

  • Regularization Techniques: Supports popular regularization techniques like L1 and L2 to reduce overfitting by minimizing parameter complexity.

  • Save and Load Models: Easily save trained models to disk and load them for future use.

Internal Usage

For use in your own Java projects, simply import the NeuralNetwork.java class file and it will immediately be usable. The following section covers the proper syntax for

  1. Initialize the Neural Network: Create a neural network by specifying the topology (number of neurons in each layer) and activation functions.

    int[] topology = {inputSize, hiddenLayerSize, outputSize};
    String[] activations = {"linear", "relu", "softmax"};
    NeuralNetwork network = new NeuralNetwork(topology, activations);
    network.Init(0.1); //initializes weights and biases according to spread amount
  2. Training: Train the neural network using your train and test/validation datasets and desired hyperparameters, taking advantage of multiple CPU cores to speed up training time.

    double[][] trainInputs = {...};
    double[][] trainOutputs = {...};
    
    double[][] testInputs = {...};
    double[][] testOutputs = {...};
    int epochs = 100;
    double learningRate = 0.01;
    int batchSize = 32;
    String lossFunction = "mse"; // or "sse" or "categorical_crossentropy"
    double decay = 0.1; // Learning rate decay
    double momentum = 0.9;
    network.clipThreshold = 1; //default gradient clipping threshold
    //set regularization of network
    network.SetRegularizationType(NeuralNetwork.RegularizationType.L2); 
    network.SetRegularizationLambda(0.001);
    NeuralNetwork.Optimizer optimizer = new NeuralNetwork.OptimizerType.Adam(0.9, 0.999); //specify optimizer for training
    //train the network with no callback
    network.Train(trainInputs, trainOutputs, testInputs, testOutputs, epochs, learningRate, batchSize, lossFunction, decay, optimizer, null);

    optimizers are of type NeuralNetwork.Optimizer and included optimizers are found in NeuralNetwork.OptimizerType. Included optimizers are:

    • SGD()

    • SGDMomentum(double momentum)

    • AdaGrad()

    • RMSProp(double decayRate)

    • Adam(double beta1, double beta2)

    Optionally, provide a custom training callback by passing in a class implementing the static NeuralNetwork.TrainingCallback interface as an argument. This can be used to make your own custom train addons like a graph visualization of the data. The ChartUpdater class has been provided to visualize accuracy data using this callback interface.

    public class Callback implements NeuralNetwork.TrainingCallback {
    	//testAccuracy is -1 if the current mini-batch doesn't have a test accuracy
    	@Override
    	public void onEpochUpdate(int epoch, int batch, double progress, double trainAccuracy, double testAccuracy) {
    		System.out.println("this statement is run for every mini-batch in training");
    	}
    }
    
    class Main {
    	public static void main(String[] args) {
    		...
    		Callback callback = new Callback();
    		network.Train(inputs, outputs, epochs, learningRate, batchSize, lossFunction, decay, optimizer, callback);
    	}
    }

    Also, custom optimizers can be made by creating a class implementing the static NeuralNetwork.Optimizer interface. This can be used to create other optimizers not already included in the NeuralNetwork class.

    public static class CustomOptimizer implements Optimizer {
    	//assign to the elements of biases and weights in the step function
    	private double[][] biases;
    	private double[][][] weights;
    	private int[] neuronsPerLayer;
    
    	@Override
    	public void initialize(int[] neuronsPerLayer, double[][] biases, double[][][] weights) {
    		this.biases = biases;
    		this.weights = weights;
    		this.neuronsPerLayer = neuronsPerLayer;
    		//other initializations
    		...
    	}
    
    	@Override
    	public void step(double[][] avgBiasGradient, double[][][] avgWeightGradient, double learningRate) {
    		for (int i = 1; i < neuronsPerLayer.length; i++) {
    			for (int j = 0; j < neuronsPerLayer[i]; j++) {
    				//set biases
    				biases[i][j] = ...
    				for (int k = 0; k < neuronsPerLayer[i - 1]; k++) {
    					//set weights
    					weights[i][j][k] = ...
    				}
    			}
    		}
    	}
    }
  3. Mutation: Mutate the neural network for a genetic algorithm (evolution).

    network.Mutate(c, v); //mutates the network with chance c and variation v
  4. Evaluation: Use the trained model to make predictions and evaluate the cost

    double[] input = {...};
    double[] prediction = network.Evaluate(input);
      
    double[] expected = {...};
    String lossFunction = "mse"; // or "sse" or "categorical_crossentropy"
    double cost = network.Cost(prediction, expected, lossFunction);
  5. Save and Load: Save the trained model to disk and load it for future use, either as a java object, which isn't human readable and doesn't transfer between programming languages but is faster, or a plain text file containing parameters, which is human readable and also transferrable between programming languages.

    // Save the model as a java object
    NeuralNetwork.Save(network, "my_model_java.nn");
    // Load the model from a file formatted as a java object
    NeuralNetwork loadedNetwork = NeuralNetwork.Load("my_model_java.nn");
    
    // Save the model as a plain text file
    NeuralNetwork.SaveParameters(network, "my_model.txt");
    // Load the model from a plain text file
    NetworkNetwork loadedTxtNetwork = NeuralNetwork.Load("my_model.txt");

    The plain text file is separated into lines that each contain a unique set of parameters specified by the first token and followed by the corresponding values, all separated by spaces. For example, one line could contain: topology 784 512 10, which would translate to a neural network with 3 layers of those sizes. The headings and their specifications are as follows:

    • numlayers: contains an integer for the number of layers in the network. usually the first line.

    • topology: contains numlayers integers describing the number of neurons in each layer. usually the second line.

    • activations: contains numlayers strings describing the activation functions for each layer. This includes the input layer, even though it is never used.

    • regularization: contains an all-caps string describing the mode of regularization and a decimal for the lambda value (regularization strength)

    • biases: contains all the biases for all the layers. in order from input to output layer, first neuron to last neuron for each layer. includes the input layer, even though it is never used.

    • weights: contains all the weights. Internally, weights is represented as a 3D array with 1st dimension layer, 2nd dimension neuron #, and 3rd dimension incoming neuron # from previous layer. All weights are flattened into series in order from input to output layer, first neuron to last neuron for each layer, and first neuron to last neuron for each previous layer.

  6. Access/Modify Parameters: get/set the parameters and information about the network.

    int numLayers = network.numLayers;
    int[] topology = network.GetTopology(); //topology[i] is # of neurons of layer i
    	 
    double[][][] weights = network.GetWeights();
    network.SetWeight(L,n2,n1,w); //sets the weight between layer L neuron n2 and layer L-1 neuron n1 to w
    
    double[][] biases = network.GetBiases();
    network.SetBias(L,n,b); //sets the bias of layer L neuron n to b
    
    String activations = network.GetActivations();
    network.SetActivation(L,act); //sets the activation of layer L to act
    
    double[][] neurons = network.GetNeurons();
    
    String info = network.toString();
    //the following two lines do the same thing
    System.out.println(info);
    System.out.println(network);

Program Usage

To use this neural network implementation, you can interact with a custom console provided by the program. Follow these steps to get started:

  1. Compile the Code: First, make sure you are working in the project directory. If you are running the full project with the console interface, run the following commands to compile and run the program:

    Unix (Mac/Linux) users:

    Compile:

    javac -cp ".:./libraries/jfreechart-1.5.3.jar" Main.java

    Run:

    java -cp ".:./libraries/jfreechart-1.5.3.jar" Main

    Windows users:

    Compile:

    javac -cp ".;./libraries/jfreechart-1.5.3.jar" Main.java

    Run:

    java -cp ".;./libraries/jfreechart-1.5.3.jar" Main

    Or, if you are just using the NeuralNetwork class, the jfreechart library can be excluded, simplifying the commands to:

    Compile:

    javac Main.java

    Run:

    java Main

This will launch the program's custom console, allowing you to control and modify neural networks.

Available Commands:

  • help: Display a list of available commands and their descriptions.

  • create: Create a new neural network by specifying its topology and activation functions.

  • init: initializes the neural network parameters (weights and biases) with the specified bias spread and weight initialization method. ('he' or 'xavier')

  • load: Load a pre-trained neural network from a saved model file.

  • train: Train the current neural network using your dataset, loss function, optimizer, and desired hyperparameters.

  • evaluate: Use the trained neural network to make predictions on new data.

  • mutate: Mutate the parameters of the network for a genetic algorithm/implementation

  • mnist: Initialize/import the MNIST dataset for use in training/evaluating or draw your own handwritten digits for the network to evaluate in realtime.

  • info: Display information about the neural network's parameters

  • reset: Reset the current neural network to default

  • modify: changes parameters of neural network

  • regularization: changes regularization type (L1, L2, none) and lambda (strength) of neural network

  • magnitude: Display information about the magnitudes of the network's parameters (min/max/average)

  • cost: Calculate the cost/accuracy of the network on a test dataset

  • save: Save the current neural network to a file for future use.

  • exit: Exit the program.

Using the Commands:

  • Type a command and press Enter to execute it. Follow the prompts to provide the required information.

  • For example, to create a new neural network, you would type create, and then follow the prompts to specify the topology and activation functions.

  • To train a network, use the train command and provide details like the path to the training file or specify mnist, number of epochs, learning rate, mini-batch size, loss function, optimizer, learning decay rate

  • For evaluation, use the evaluate command, and input the data you want to predict on, or mnist [case #]

  • The mnist dataset has been built into the program to allow for building/evaluating hand-drawn digit recognition networks directly using the console and recieving a realtime accuracy visualization to assist with training. You can run mnist test to draw your own hand-written digits for the network to evaluate in realtime. the mnist command to import the MNIST dataset has an optional augmented modifier to randomly apply affine transformations to the data, making the network more generalizable at the cost of some accuracy and convergence speed. mnist networks MUST HAVE input size 784 and output size 10. In the majority of cases, the output layer has softmax activation and is trained using the categorical_crossentropy loss function.

Save and Load Models:

  • You can save the trained model using the save command. This will save the model to a file for future use.

  • To load a pre-trained model, use the load command and specify the file path of the saved model.

Training/Test Data File Formatting:

All training and test dataset files must be formatted in the following way:

  • first line has 3 numbers specifying # cases, input and output size
  • every line is a separate training case
  • on each line, input is separated by spaces, then equal sign, then output separated by spaces

Exiting the Program:

  • To exit the program, simply type exit, and the program will terminate.

Examples

A few neural networks and their training sets have been pre-included into the project, ready to be loaded in. They are:

  • SavedNetwork1: simple neural network to add 2 numbers (object mode)
  • SavedNetwork2: deep neural network to add 2 numbers (object mode)
  • TrainSet1: training/test dataset for adding 2 numbers (can be used for SavedNetwork1 and SavedNetwork2) (object mode)
  • MNISTNetwork: an untrained neural network with the correct topology to evaluate MNIST cases (digit recognition). accuracy ≈ 10.61% (object mode)
  • MNISTNetworkTrained: a trained neural network that evaluates MNIST cases (digit recognition) with high, generalized accuracy from training on an augmented data set. Good for testing your own digits. accuracy ≈ 98.14% (object mode)
  • MNISTParams: same as MNISTNetworkTrained, but as plain text (parameters mode)
  • MNISTNetworkOverfitted: a neural network that has been trained excessively, making it overfit to the MNIST dataset. Bad for testing your own digits. accuracy ≈ 99.94% (object mode)

About

flexible and extensible implementation of a multithreaded feedforward neural network in Java including popular optimizers, wrapped up in a console user interface

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published