Differences between revisions 11 and 12
Revision 11 as of 2011-02-22 12:37:49
Size: 2113
Editor: dyn-176108
Comment:
Revision 12 as of 2011-02-22 14:16:26
Size: 4235
Editor: JunHu
Comment:
Deletions are marked like this. Additions are marked like this.
Line 21: Line 21:

== Examples ==
{{!java
import org.neuroph.core.NeuralNetwork;
import org.neuroph.nnet.MultiLayerPerceptron;
import org.neuroph.core.learning.TrainingSet;
import org.neuroph.core.learning.TrainingElement;
import org.neuroph.core.learning.SupervisedTrainingElement;
import java.util.Vector;
import org.neuroph.util.TransferFunctionType;

/**
 * This sample shows how to create, train, save and load simple Multi Layer Perceptron
 */

void setup() {

    // create training set (logical XOR function)
    TrainingSet trainingSet = new TrainingSet();
    trainingSet.addElement(new SupervisedTrainingElement(new double[] {
      0, 0
    }
    , new double[] {
      0
    }
    ));
    trainingSet.addElement(new SupervisedTrainingElement(new double[] {
      0, 1
    }
    , new double[] {
      1
    }
    ));
    trainingSet.addElement(new SupervisedTrainingElement(new double[] {
      1, 0
    }
    , new double[] {
      1
    }
    ));
    trainingSet.addElement(new SupervisedTrainingElement(new double[] {
      1, 1
    }
    , new double[] {
      0
    }
    ));

    // create multi layer perceptron
    MultiLayerPerceptron myMlPerceptron = new MultiLayerPerceptron(TransferFunctionType.TANH, 2, 3, 1);
    // learn the training set
    myMlPerceptron.learnInSameThread(trainingSet);

    // test perceptron
    System.out.println("Testing trained neural network");
    testNeuralNetwork(myMlPerceptron, trainingSet);

    // save trained neural network
    myMlPerceptron.save("myMlPerceptron.nnet");

    // load saved neural network
    NeuralNetwork loadedMlPerceptron = NeuralNetwork.load("myMlPerceptron.nnet");

    // test loaded neural network
    System.out.println("Testing loaded neural network");
    testNeuralNetwork(loadedMlPerceptron, trainingSet);
    
    noLoop();
  }

void testNeuralNetwork(NeuralNetwork nnet, TrainingSet tset) {
  
    nnet.setInput(new double[]{1, 0});
    nnet.calculate();
    double networkOutput = nnet.getOutput()[0];
    System.out.println(" Output: " + networkOutput);
  }


}}

Preparation

  • Follow the instructions from the assignment Creative Programming to prepare the SoftwareEnvironment, including Processing, Arduino and the AdMoVeo Robot. If you are not planning to use the GUI components from the ControlP5 library, you can skip the ControlP5 part.

  • Install Neuroph, a lightweight Java neural network framework.

    • Download Neuroph Studio and install it.

    • Download neuroph-2.5b.zip. Unzip it into C:\Programs. In C:\Programs\neuroph-2.5b you shall be able to find neuroph-2.5b.jar. (You can unzip it to anywhere you want, but then remember where it is for later reference).

    • In "Processing Sketchbook location"\libraries, create a sub-directory "neuroph".
    • In "Processing Sketchbook location"\libraries\neuroph, create a sub-directory "library".
    • From C:\Programs\neuroph-2.5b copy neuroph-2.5b.jar to "Processing Sketchbook location"\libraries\neuroph\library. Rename neuroph-2.5b.jar to neuroph.jar.
    • From C:\Programs\neuroph-2.5b\lib copy two jar files found there to "Processing Sketchbook location"\libraries\neuroph\library.
  • If you are confident enough, you can also try to use Eclipse to program in Java. Then try to follow the instructions from the assignment Processing2Java. Remember you shall include the aforementioned jar files in the projects if you want to use neuroph.

References

Examples

{{!java import org.neuroph.core.NeuralNetwork; import org.neuroph.nnet.MultiLayerPerceptron; import org.neuroph.core.learning.TrainingSet; import org.neuroph.core.learning.TrainingElement; import org.neuroph.core.learning.SupervisedTrainingElement; import java.util.Vector; import org.neuroph.util.TransferFunctionType;

/**

  • This sample shows how to create, train, save and load simple Multi Layer Perceptron

    */

void setup() {

  • // create training set (logical XOR function)

    TrainingSet trainingSet = new TrainingSet(); trainingSet.addElement(new SupervisedTrainingElement(new double[] {

    • 0, 0
    } , new double[] {
    • 0
    } ));

    trainingSet.addElement(new SupervisedTrainingElement(new double[] {

    • 0, 1
    } , new double[] {
    • 1
    } ));

    trainingSet.addElement(new SupervisedTrainingElement(new double[] {

    • 1, 0
    } , new double[] {
    • 1
    } ));

    trainingSet.addElement(new SupervisedTrainingElement(new double[] {

    • 1, 1
    } , new double[] {
    • 0
    } )); // create multi layer perceptron

    MultiLayerPerceptron myMlPerceptron = new MultiLayerPerceptron(TransferFunctionType.TANH, 2, 3, 1); // learn the training set myMlPerceptron.learnInSameThread(trainingSet); // test perceptron System.out.println("Testing trained neural network"); testNeuralNetwork(myMlPerceptron, trainingSet); // save trained neural network myMlPerceptron.save("myMlPerceptron.nnet"); // load saved neural network

    NeuralNetwork loadedMlPerceptron = NeuralNetwork.load("myMlPerceptron.nnet"); // test loaded neural network System.out.println("Testing loaded neural network"); testNeuralNetwork(loadedMlPerceptron, trainingSet); noLoop();

  • }

void testNeuralNetwork(NeuralNetwork nnet, TrainingSet tset) {

  • nnet.setInput(new double[]{1, 0}); nnet.calculate(); double networkOutput = nnet.getOutput()[0]; System.out.println(" Output: " + networkOutput);
  • }

}}

JunHu: NeurophNeuralNetwork (last edited 2016-12-12 20:50:49 by JunHu)