Convert examples to all-in-one networks#4458
Conversation
|
Additional con for this implementation:
See discussion at #3864 - changing the network to require fewer flags makes them less legible. |
Tested: train_lenet.sh
5cebb09 to
bc6e6f1
Compare
Tested: train_quick.sh train_full.sh
Tested: train_mnist_siamese.sh
bc6e6f1 to
8219866
Compare
|
Some of the documentation isn't really updated yet. I'll wait for feedback on whether this type of change is desired before diving into the docs. |
|
I hope these changes are merged. I personally prefer the more explicit method of including the phase and the stage. However, I don't think that is a reason to hold back merging these changes. |
|
@williford By "more explicit" do you mean this: instead of this: I'm happy with that if you think it's clearer. @shelhamer do you have an opinion on this? Is there any chance this gets merged? |
|
@lukeyeager exactly. I wouldn't make the change unless others agree. I might just be biased because I'm used to having the phase. |
|
Closing after a year of inactivity. This PR branch is backed up at https://github.com/lukeyeager/caffe/tree/backup/all-in-one-examples. |
|
Hello, After spending more than one day, I have not still be able to run any model related to resnet or densenet on DIGITS6. The only thing I can do is training google lenet (that is shipped with Digits6) I am taking any network from some sources (i.e) train_val.prototxt s. However, I can not train them. I ll drive crazy. There is no good documentation about how to make it possible to train those nets. Even the alexnet is problematic (shipped with digits6 model store) I can not figure out why you did something like that. If I cannot run it, what is the purpose of Digits 6? Everybody who uses digits aims to gain time and play with the networks. With this version, it seems that it ll not be possible to train them with in 2 minutes. Now, I want to ask @lukeyeager . Please convert the https://github.com/antingshen/resnet-protofiles/blob/master/ResNet_50_train_val.prototxt in order to be trained in Digits6. I have done every possible combination. It still says: "ERROR: Layer 'loss' references bottom 'label' at the TEST stage however this blob is not included at that stage. Please consider using an include directive to limit the scope of this layer." I am excluding, including but nothing works. Either I am blind or your software is buggy. I even checked the caffe_train.py in order to understand whether or not it is out dated. But it is new. So please clarify us. I see lots of people who suffers from this error. Did we install it improperly or is there a bug? Or, must I convert every single layer to make it appropriate all-in-one network? Btw, as I stated, I have tried almost every possible combination including silly ones. But it does NOT work! Looking forward to hear from you soon. |
Implementation of #3864.
Pros
.prototxtto read since the number of files decreasesCons
TODO: