Deep Learning Step5 Back Propagation and Composition Function
Deep Learning Step5 Step 5 Back Propagation Why We Need Back Propagation? In supervised learning systems, Back Propagation gets minimum value of loss function. Loss function is used to decide optimized hyper parameters which will be described later. Differentiation of Composition Function First, we need to know how to do differentiation of composition. For example, the following functions are ready. Input Data : x Function A(), B() ,C() Output each : a = A(X) , b = B(a) , c = C(b) Fig.5.1 Forward Propagation If x is put into A(), you will get a. As well as a is put into B and b into C. This way is called Forward Propagation(Fig.5.1). If we want value of dC(b)/dx ( c = C(b) = C(B(A(X))) => dc/dx = dC(b)/dx = dC(B(A(X)))/dx) , we can solve this equation like the below. Fig.5.2 Backward Propagation This data flow in Fig.5.2 is called back propagation because the value of differentiation of dc/dc is g...