@@ -42,23 +42,23 @@ def __init__(self, num_layers_in_fc_layers = 1024):
4242 nn .ReLU (inplace = True ),
4343 nn .MaxPool2d (kernel_size = (3 ,3 ), stride = (2 ,2 )),
4444
45- nn .Conv2d (256 , 2048 , kernel_size = (5 ,4 ), padding = (0 ,0 )),
46- nn .BatchNorm2d (2048 ),
45+ nn .Conv2d (256 , 512 , kernel_size = (5 ,4 ), padding = (0 ,0 )),
46+ nn .BatchNorm2d (512 ),
4747 nn .ReLU (),
4848 );
4949
5050 self .netfcaud = nn .Sequential (
51- nn .Linear (2048 , 2048 ),
52- nn .BatchNorm1d (2048 ),
51+ nn .Linear (512 , 512 ),
52+ nn .BatchNorm1d (512 ),
5353 nn .ReLU (),
54- nn .Linear (2048 , num_layers_in_fc_layers ),
54+ nn .Linear (512 , num_layers_in_fc_layers ),
5555 );
5656
5757 self .netfclip = nn .Sequential (
58- nn .Linear (2048 , 2048 ),
59- nn .BatchNorm1d (2048 ),
58+ nn .Linear (512 , 512 ),
59+ nn .BatchNorm1d (512 ),
6060 nn .ReLU (),
61- nn .Linear (2048 , num_layers_in_fc_layers ),
61+ nn .Linear (512 , num_layers_in_fc_layers ),
6262 );
6363
6464 self .netcnnlip = nn .Sequential (
@@ -85,8 +85,8 @@ def __init__(self, num_layers_in_fc_layers = 1024):
8585 nn .ReLU (inplace = True ),
8686 nn .MaxPool3d (kernel_size = (1 ,3 ,3 ), stride = (1 ,2 ,2 )),
8787
88- nn .Conv3d (256 , 2048 , kernel_size = (1 ,6 ,6 ), padding = 0 ),
89- nn .BatchNorm3d (2048 ),
88+ nn .Conv3d (256 , 512 , kernel_size = (1 ,6 ,6 ), padding = 0 ),
89+ nn .BatchNorm3d (512 ),
9090 nn .ReLU (inplace = True ),
9191 );
9292
0 commit comments