- neuro_src.zip
- neuro_src
- Docs
- AForge.Core.chm
- AForge.Neuro.chm
- Release
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Samples
- Neuro
- Back Propagation
- Approximation
- TimeSeries
- XORProblem
- Simple
- Delta Rule Learning
- One-Layer Perceptron Classifier
- Perceptron Classifier
- SOM
- Sources
- neuro_demo.zip
- neuro_demo
- Back Propagation
- Approximation
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Approximation.exe
- Data Samples
- TimeSeries
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Data Samples
- TimeSeries.exe
- XORProblem
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- XORProblem.exe
- Simple
- Delta Rule Learning
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Classifier.exe
- Data Samples
- One-Layer Perceptron Classifier
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Classifier.exe
- Data Samples
- Perceptron Classifier
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- Classifier.exe
- Data Samples
- SOM
- 2DOrganizing
- 2DOrganizing.exe
- AForge.dll
- AForge.Neuro.dll
- Color
- AForge.dll
- AForge.Neuro.dll
- Color.exe
- TSP
- AForge.Controls.dll
- AForge.dll
- AForge.Neuro.dll
- TSP.exe
|
// AForge Neural Net Library
//
// Copyright � Andrew Kirillov, 2005-2006
// andrew.kirillov@gmail.com
//
namespace AForge.Neuro.Learning
{
using System;
/// <summary>
/// Kohonen Self Organizing Map (SOM) learning algorithm
/// </summary>
///
/// <remarks>This class implements Kohonen's SOM learning algorithm and
/// is widel� used in clusterization tasks. The class allows to train
/// <see cref="DistanceNetwork">Distance Networks</see>.</remarks>
///
public class SOMLearning : IUnsupervisedLearning
{
// neural network to train
private DistanceNetwork network;
// network's dimension
private int width;
private int height;
// learning rate
private double learningRate = 0.1;
// learning radius
private double learningRadius = 7;
// squared learning radius multiplied by 2 (precalculated value to speed up computations)
private double squaredRadius2 = 2 * 7 * 7;
/// <summary>
/// Learning rate
/// </summary>
///
/// <remarks>Determines speed of learning. Value range is [0, 1].
/// Default value equals to 0.1.</remarks>
///
public double LearningRate
{
get { return learningRate; }
set
{
learningRate = Math.Max( 0.0, Math.Min( 1.0, value ) );
}
}
/// <summary>
/// Learning radius
/// </summary>
///
/// <remarks>Determines the amount of neurons to be updated around
/// winner neuron. Neurons, which are in the circle of specified radius,
/// are updated during the learning procedure. Neurons, which are closer
/// to the winner neuron, get more update.<br /><br />
/// Default value equals to 7.</remarks>
///
public double LearningRadius
{
get { return learningRadius; }
set
{
learningRadius = Math.Max( 0, value );
squaredRadius2 = 2 * learningRadius * learningRadius;
}
}
/// <summary>
/// Initializes a new instance of the <see cref="SOMLearning"/> class
/// </summary>
///
/// <param name="network">Neural network to train</param>
///
/// <remarks>This constructor supposes that a square network will be passed for training -
/// it should be possible to get square root of network's neurons amount.</remarks>
///
public SOMLearning( DistanceNetwork network )
{
// network's dimension was not specified, let's try to guess
int neuronsCount = network[0].NeuronsCount;
width = (int) Math.Sqrt( neuronsCount );
if ( width * width != neuronsCount )
{
throw new ArgumentException( "Invalid network size" );
}
// ok, we got it
this.network = network;
this.width = width;
this.height = height;
}
/// <summary>
/// Initializes a new instance of the <see cref="SOMLearning"/> class
/// </summary>
///
/// <param name="network">Neural network to train</param>
/// <param name="width">Neural network's width</param>
/// <param name="height">Neural network's height</param>
///
/// <remarks>The constructor allows to pass network of arbitrary rectangular shape.
/// The amount of neurons in the network should be equal to <b>width</b> * <b>height</b>.
/// </remarks>
///
public SOMLearning( DistanceNetwork network, int width, int height )
{
// check network size
if ( network[0].NeuronsCount != width * height )
{
throw new ArgumentException( "Invalid network size" );
}
this.network = network;
this.width = width;
this.height = height;
}
/// <summary>
/// Runs learning iteration
/// </summary>
///
/// <param name="input">input vector</param>
///
/// <returns>Returns learning error - summary absolute difference between updated
/// weights and according inputs. The difference is measured according to the neurons
/// distance to the winner neuron.</returns>
///
public double Run( double[] input )
{
double error = 0.0;
// compute the network
network.Compute( input );
int winner = network.GetWinner( );
// get layer of the network
Layer layer = network[0];
// check learning radius
if ( learningRadius == 0 )
{
Neuron neuron = layer[winner];
// update weight of the winner only
for ( int i = 0, n = neuron.InputsCount; i < n; i++ )
{
neuron[i] += ( input[i] - neuron[i] ) * learningRate;
}
}
else
{
// winner's X and Y
int wx = winner % width;
int wy = winner / width;
// walk through all neurons of the layer
for ( int j = 0, m = layer.NeuronsCount; j < m; j++ )
{
Neuron neuron = layer[j];
int dx = ( j % width ) - wx;
int dy = ( j / width ) - wy;
// update factor ( Gaussian based )
double factor = Math.Exp( - (double) ( dx * dx + dy * dy ) / squaredRadius2 );
// update weight of the neuron
for ( int i = 0, n = neuron.InputsCount; i < n; i++ )
{
// calculate the error
double e = ( input[i] - neuron[i] ) * factor;
error += Math.Abs( e );
// update weight
neuron[i] += e * learningRate;
}
}
}
return error;
}
/// <summary>
/// Runs learning epoch
/// </summary>
///
/// <param name="input">array of input vectors</param>
///
/// <returns>Returns summary learning error for the epoch. See <see cref="Run"/>
/// method for details about learning error calculation.</returns>
///
public double RunEpoch( double[][] input )
{
double error = 0.0;
// walk through all training samples
foreach ( double[] sample in input )
{
error += Run( sample );
}
// return summary error
return error;
}
}
}
|
By viewing downloads associated with this article you agree to the Terms of Service and the article's licence.
If a file you wish to view isn't highlighted, and is a text file (not binary), please
let us know and we'll add colourisation support for it.
Started software development at about 15 years old and it seems like now it lasts most part of my life. Fortunately did not spend too much time with Z80 and BK0010 and switched to 8086 and further. Similar with programming languages – luckily managed to get away from BASIC and Pascal to things like Assembler, C, C++ and then C#. Apart from daily programming for food, do it also for hobby, where mostly enjoy areas like Computer Vision, Robotics and AI. This led to some open source stuff like
AForge.NET,
Computer Vision Sandbox,
cam2web,
ANNT, etc.