(Review cs231n) Optimized Methods

Mini-batch SGD的步驟: 1.Sample a batch of data 2.Forward prop it through the graph,get loss 3.backprop to calculate the gradient 4. updata the parameters using the gradient The initialization of weights
相關文章
相關標籤/搜索