AI-Perceptron
view release on metacpan or search on metacpan
Gradient-Descent* model.
MODEL
Model of a Perceptron
+---------------+
X[1] o------ |W[1] T |
X[2] o------ |W[2] +---------+ +-------------------+
. | . | ___ |_________| __ Squarewave |_______\ Output
. | . | \ | S | __| Generator | /
. | . | /__ | +-------------------+
X[n] o------ |W[n] | Sum |
+-----+---------+
S = T + Sum( W[i]*X[i] ) as i goes from 1 -> n
Output = 1 if S > 0; else -1
Where "X[n]" are the perceptron's *inputs*, "W[n]" are the *Weights*
that get applied to the corresponding input, and "T" is the *Threshold*.
The *squarewave generator* just turns the result into a positive or
negative number.
So in summary, when you feed the perceptron some numeric inputs you get
either a positive or negative output depending on the input's weights
and a threshold.
TRAINING
Usually you have to train a perceptron before it will give you the
outputs you expect. This is done by giving the perceptron a set of
examples containing the output you want for some given inputs:
-1 => -1, -1
-1 => 1, -1
-1 => -1, 1
1 => 1, 1
If you've ever studied boolean logic, you should recognize that as the
truth table for an "AND" gate (ok so we're using -1 instead of the
commonly used 0, same thing really).
You *train* the perceptron by iterating over the examples and adjusting
the *weights* and *threshold* by some value until the perceptron's
output matches the expected output of each example:
while some examples are incorrectly classified
update weights for each example that fails
The value each weight is adjusted by is calculated as follows:
delta[i] = learning_rate * (expected_output - output) * input[i]
Which is know as a negative feedback loop - it uses the current output
as an input to determine what the next output will be.
Also, note that this means you can get stuck in an infinite loop. It's
not a bad idea to set the maximum number of iterations to prevent that.
CONSTRUCTOR
new( [%args] )
Creates a new perceptron with the following default properties:
num_inputs = 1
learning_rate = 0.01
threshold = 0.0
weights = empty list
Ideally you should use the accessors to set the properties, but for
backwards compatability you can still use the following arguments:
Inputs => $number_of_inputs (positive int)
N => $learning_rate (float)
W => [ @weights ] (floats)
The number of elements in *W* must be equal to the number of inputs
plus one. This is because older version of AI::Perceptron combined
the threshold and the weights a single list where W[0] was the
threshold and W[1] was the first weight. Great idea, eh? :) That's
why it's *DEPRECATED*.
ACCESSORS
num_inputs( [ $int ] )
Set/get the perceptron's number of inputs.
learning_rate( [ $float ] )
Set/get the perceptron's number of inputs.
weights( [ \@weights ] )
Set/get the perceptron's weights (floats).
For backwards compatability, returns a list containing the
*threshold* as the first element in list context:
($threshold, @weights) = $p->weights;
This usage is *DEPRECATED*.
threshold( [ $float ] )
Set/get the perceptron's number of inputs.
training_examples( [ \@examples ] )
Set/get the perceptron's list of training examples. This should be a
list of arrayrefs of the form:
[ $expected_result => @inputs ]
max_iterations( [ $int ] )
Set/get the perceptron's number of inputs, a negative value implies
no maximum.
METHODS
compute_output( @inputs )
Computes and returns the perceptron's output (either -1 or 1) for
the given inputs. See the above model for more details.
add_examples( @training_examples )
Adds the @training_examples to to current list of examples. See
training_examples() for more details.
train( [ @training_examples ] )
Uses the *Stochastic Approximation of the Gradient-Descent* model to
( run in 0.565 second using v1.01-cache-2.11-cpan-39bf76dae61 )