Conversation
|
Looks good. Thanks Evan! |
|
@jeffdonahue could you decide what to do with tnarihi/caffe@7d45526 too since you and Takuya talked about the gradient checker? Thanks. |
|
LGTM, thanks @tnarihi i and @shelhamer. I'm also not sure of the best way to go with changing the gradient checker -- always checking accumulation is safer, but anyone who's developed a layer of their own with parameters may be confused as to why their implementation is suddenly failing the gradient checker. I guess ideally there would be a first pass that does the current gradient check and, only if that passes, also checks accumulation afterwards and fails with an error message pointing to failed accumulation if applicable. Seems like the code would get kind of ugly though... |
Adapt the PReLU gradient computation to accumulation now that #1977 is in.
@tnarihi could you double-check this? I cherry-picked your commit and cleaned up comments. Thanks.