Advertisement

Help me with my back propagation formula

Started by February 04, 2006 05:04 AM
-1 comments, last by Mizipzor 18 years, 9 months ago
I still havent been able to crack my ANN problem. Instead of posting source and ask for help, I ask if anyone could tell me if I got the maths right or not. First off, heres how I describe different parts of the net: w(l,f,n) - Weight for input f to neuron n in layer l. x(l,n) - Output from neuron n in layer l. l=0 is inputs from envoirment and n=0 is the bias. d(l,n) - Deltas for neuron n in layer l. Used in the back propagation function. Now, heres the layout of my net:

Input	Hidden_1		Hidden_2		Output_Neuron

x(0,0)			x(1,0)			x(2,0)			<- Biases
			
x(0,1)	w(0,0,0)	x(1,1)	w(1,0,0)	x(2,1)	w(2,0,0)
x(0,2)	w(0,1,0)		w(1,1,0)		w(2,1,0)	x(3,0)	// Total net output
	w(0,2,0)		w(1,2,0)		w(2,2,0)
	
	w(0,0,1)	x(1,2)	w(1,0,1)	x(2,2)
	w(0,1,1)		w(1,1,1)
	w(0,2,1)		w(1,2,1)

			d(0,n)			d(1,n)			d(2,n)	// Deltas for each layer



I calculate the delta for the output neuron with this forula: // output layer delta, we only have one output // d(2,0) = x(3,0)(1 - x(3,0))(d - x(3,0)) Deltas for the hidden layers are calculated like this: // d(l,n) = x(l,n) (1 - x(l,n)) w(l+1,n,n-1) d(l+1,n) Now I alter the weights, thats done with this formula (h is learning constant): // Deltas calculated, now alter the weights // w(l,f,n) = h * x(l,f) * d(l,n) I have gone through the code several times, I cannot find anything wrong. So Ive made the conclusion that I must have missunderstood the maths. Could anyone tell me if these formulas would work in a neural net? If the math is correct, at least I know the error is somewhere in the code. Anyone, please^2 help me out with the math part. Thanks in advance. [smile]

This topic is closed to new replies.

Advertisement