How do I solve $\frac{\partial}{\partial w_{jk}} { \sum_j w_{jk} . o_j }$
Please I want to know why $\frac{\partial}{\partial W_{jk}} { \sum_j W_{jk} . o_j }$ = $o_j$
I know $\frac{\partial}{\partial W_{jk}} { W_{jk} . o_j }$ = $o_j$ but I got confused by the summation symbol.
I was reading it from "MAKE YOUR OWN NEURAL NETWORK" book (page 96) by TARIQ RASHIQ. $W_{jk}$ represent the weight connecting node j to node k. And $o_j$ represent the output of node j
But then what I was thinking was: if the partial differentiation of $W_{jk}$ with respect to $W_{jk}$ is one , then it means we are getting $\sum1 . oj$ and $\sum1$ is 1. So finally we will get $o_j$
Solution 1:
Please I want to know why $\frac{\partial}{\partial W_{jk}} { \sum_j W_{jk} . o_j } = o_j$
...but I got confused by the summation symbol.
That is not unusual. You have overloaded the $j$ token; one use is as a free token, and the other is bound to the summation. Although this is legal, it is often a source of confusion. For clarity you should use a fresh token in the series. Recall that $\sum_j a_j=\sum_i a_i$ .
Then we may partition the series, extracting the one term where the series index equals the free token from all the others, where they do not match.
$$\dfrac{\partial\quad}{\partial W_{jk}} { \sum_i W_{ik} \cdot o_i } ~=~ \dfrac{\partial\quad}{\partial W_{jk}} ( W_{jk} \cdot o_j )+\sum_{i~:~i\neq j}\dfrac{\partial\quad}{\partial W_{jk}}(W_{ik}\cdot o_i)$$
You can handle the first term, while the remainder vanish.