10aima/learning/statistics/StandardBackPropogationjava/lang/Object4aima/learning/statistics/NeuralNetworkTrainingSchemeneuronDeltaMapLjava/util/Hashtable; SignatureJLjava/util/Hashtable; neuronBiasMap linkWeightMapHLjava/util/Hashtable; learningRateD()VCode  java/util/Hashtable       ?   LineNumberTableLocalVariableTablethis2Laima/learning/statistics/StandardBackPropogation; backPropogateP(Laima/learning/statistics/FeedForwardNetwork;Ljava/util/List;Ljava/util/List;)Vx(Laima/learning/statistics/FeedForwardNetwork;Ljava/util/List;Ljava/util/List;)V+aima/learning/statistics/FeedForwardNetwork)propogateInput(Ljava/util/List;)V +, *-calculateOutputLayerDelta@(Laima/learning/statistics/FeedForwardNetwork;Ljava/util/List;)V /0 1calculateHiddenLayersDelta0(Laima/learning/statistics/FeedForwardNetwork;)V 34 5network-Laima/learning/statistics/FeedForwardNetwork;inputLjava/util/List; correctOutputLocalVariableTypeTable$Ljava/util/List;T(Laima/learning/statistics/FeedForwardNetwork;Ljava/util/List;)VgetOutputLayer"()Laima/learning/statistics/Layer; ?@ *Aaima/learning/statistics/LayerCiterator()Ljava/util/Iterator; EF DGgetError"(Ljava/util/List;)Ljava/util/List; IJ DKjava/util/ListM NGjava/util/IteratorPnext()Ljava/lang/Object; RS QTaima/learning/statistics/NeuronVjava/lang/DoubleZ doubleValue()D \] [^valueOf(D)Ljava/lang/Double; `a [bput8(Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object; de fhasNext()Z hi Qj outputLayer Laima/learning/statistics/Layer; neuronIterLjava/util/Iterator; errorIter7Ljava/util/Iterator;(Ljava/util/Iterator;getHiddenLayers()Ljava/util/List; st *usize()I wx Nyget(I)Ljava/lang/Object; {| N} getNeurons t DoutLinks t Waima/learning/statistics/Linkweight ] target#()Laima/learning/statistics/Neuron; &(Ljava/lang/Object;)Ljava/lang/Object; { getActivationFuncton/()Laima/learning/statistics/ActivationFunction; W activation ] W+aima/learning/statistics/ActivationFunctionderiv(D)D  hiddenLayersiIlayerneuron!Laima/learning/statistics/Neuron; weightsumlLaima/learning/statistics/Link;2Ljava/util/List;delta2(Laima/learning/statistics/Layer;)Ljava/util/List;F(Laima/learning/statistics/Layer;)Ljava/util/List;java/util/ArrayList add(Ljava/lang/Object;)Z NlistnupdateWeightsAndBiases$(Laima/learning/statistics/Neuron;)V  hiddenLayerbias ] WsetBias(D)V WinLinks t Wsource  setWeight link weightChangeerror@(Ljava/util/List;Laima/learning/statistics/FeedForwardNetwork;)DT(Ljava/util/List;Laima/learning/statistics/FeedForwardNetwork;)D J *aima/util/Util sumOfSquares(Ljava/util/List;)D expectedOutput SourceFileStandardBackPropogation.java!        k-**Y*Y*Y*!"%,# -$%&' (+,.*+-2*+6" !#*$%789:;:<9=;=/0 >U+BN-H:-,LO:)*UWXU[_kcgWk kɱ"& %& '('*,@*T/#>U$%U78U;:Plm Jno>po< U;= Jnq>pr34] +vM,zd>,~D:O:sUW:9O: , U:  * [_kc9 k*kcgWkf"2 234586;7V8s7}:53>#R$%78:m8n;aV' <  ?YM+O: UWN,*-[Wk,"AB!C3B=E#*?$%?m7:!< 7=4s+BON-UWM*,-k+vON7-UDM,O:UW:*k-kƱ"& IJI(L?MYN_MiLrR#4s$%s78?3mY*+[_I*++cgW++*!(kg+¹O:?U:*cgW*!(kŶk9gk"& UVW.XHYZZk[xX^#4$%tH:k  Y ,+ϸկ"a#  $% : 78<  =