Skip to content

Commit

Permalink
fix typo, rewrite graph
Browse files Browse the repository at this point in the history
  • Loading branch information
dzhwinter committed Sep 5, 2017
1 parent b317cbf commit dbaaa49
Showing 1 changed file with 4 additions and 0 deletions.
4 changes: 4 additions & 0 deletions paddle/framework/multigpu.md
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,10 @@ These two operators need the Multi-GPU context support.

Need to notice that Allreduce operator force GPUs synchronized at that point. Every device only need runs sub-graph in a loop style forever, the whole training process in asynchronous or synchronous mode depends on the Allreduce point in the graph.

For the simplest implement, when each GPU compute the gradient of `W`, followed with a `AllReduce` operator, accumulate the `dW` to full batch of data, then run the optimize process individually and apply the gradient to its `W`.

In fact, in the way of every GPU optimized full batch of data, wasted (n-1) GPU compute resources. We will enhance it in the next stage.

### Benefits

- can easily move the optimize sub-graph to parameter server, multi-GPU feature can be compatible with distributed support design.
Expand Down

0 comments on commit dbaaa49

Please sign in to comment.