Shallow Neural Network Week 3

Single Sample

Symbols

X=(x1xnx),Y=(y1yny),
Z[l]=(z1[l]znl[l]),1lL
A[l]=(a1[l]anl[l]),A~[l]=(a0[l]a1[l]anl[l])=(1A[l]),0lL
W[l]=(wij[l])nl×nl1,w[l]=(w1,0[l]wnl,0[l]),W~[l]=(w[l]W[l]),1l<L

Neural Network Architecture

X=A[0]Z[1]A[1]Z[L]A[L]=Y^

Loss Function

zi[l]=j=0nl1wij[l]a~j[l1],1inl,1lL
Zl=W[l]A~[l1],1lL
ai[l]=g(zi[l]),1inl,1lL
A[l]=g(Z[l]),1lL
loss(X,Y)=i=1ny[yilny^i+(1yi)ln(1y^i)]

公式

zi[L]loss(X,Y)=dy^idzi[L]y^iloss(X,Y)
=g(z[L])[yi1y^i(1yi)11y^i]
=y^i(1y^i)[yi1y^i(1yi)11y^i]
=(1yi)y^iyi(1y^i)
=y^iyi,1inL

zj[l]loss(X,Y)=i=1nl+1zi[l+1]zj[l]zi[l+1]loss(X,Y)
=i=1nl+1g(zj[l])wij[l]zi[l+1]loss(X,Y)
=g(zj[l])i=1nl+1wij[l]zi[l+1]loss(X,Y),1jsl,1l<L
因此
Z[l]loss(X,Y)={A[L]Y,l=Lg(Z[l]) . ((W[l+1])Z[l+1]loss(X,Y)),1l<L
where .* is element-wise product.

wij[l]loss(X,Y)=zi[l]loss(X,Y)a~j[l1],1isl+1,0jsl,1lL
因此
W~[l]loss(X,Y)=Z[l]loss(X,Y)A~[l1],1lL

Multiple Samples

Symbols

X=(X(1),,X(m)),
Y=(Y(1),,Y(m)),
Z[l]=(Z[l](1),,Z[l](m)),1lL
A[l]=(A[l](1),,A[l](m)),0lL
A~[l]=(A~[l](1),,A~[l](m)),0lL
Z[l]=(Z[l]loss(X(1),Y(1)),,Z[l]loss(X(m),Y(m)))nl×m,1lL

Cost Function

cost(X,Y)=1mi=1mloss(X(i),Y(i))

公式

Z[l]=W[l]A~[l1],1l<L
A[l]=g(Z[l]),1lL
g(Z[l])=A[l] . (1nl×mA[l]),1lL

Z[l]={A[L]Y,l=Lg(Z[l]) . ((W[l+1])Z[l+1]),1l<L
W~[l]cost(X,Y)=1mZ[l]A~[l1],1lL

發表評論
所有評論
還沒有人評論,想成為第一個評論的人麼? 請在上方評論欄輸入並且點擊發布.
相關文章