@@ -15,7 +15,7 @@ datasets:
15
15
key : ' torchvision.transforms.RandomResizedCrop'
16
16
init :
17
17
kwargs :
18
- size : &input_size [224, 224 ]
18
+ size : &input_size [299, 299 ]
19
19
- !import_call
20
20
key : ' torchvision.transforms.RandomHorizontalFlip'
21
21
init :
@@ -44,7 +44,7 @@ datasets:
44
44
key : ' torchvision.transforms.Resize'
45
45
init :
46
46
kwargs :
47
- size : 256
47
+ size : 327
48
48
- !import_call
49
49
key : ' torchvision.transforms.CenterCrop'
50
50
init :
@@ -131,11 +131,11 @@ train:
131
131
student :
132
132
forward_proc : ' forward_batch_only'
133
133
adaptations :
134
- sequential : ['bottleneck ', 'Mixed_5b', 'Mixed_5c', 'Mixed_5d', 'Mixed_6a', 'Mixed_6b', 'Mixed_6c', 'Mixed_6d', 'Mixed_6e', 'Mixed_7a', 'Mixed_7b', 'Mixed_7c']
134
+ sequential : ['bottleneck_layer ', 'Mixed_5b', 'Mixed_5c', 'Mixed_5d', 'Mixed_6a', 'Mixed_6b', 'Mixed_6c', 'Mixed_6d', 'Mixed_6e', 'Mixed_7a', 'Mixed_7b', 'Mixed_7c']
135
135
frozen_modules : ['Mixed_5b', 'Mixed_5c', 'Mixed_5d', 'Mixed_6a', 'Mixed_6b', 'Mixed_6c', 'Mixed_6d', 'Mixed_6e', 'Mixed_7a', 'Mixed_7b', 'Mixed_7c']
136
136
forward_hook :
137
137
input : []
138
- output : ['bottleneck ', 'Mixed_5d', 'Mixed_6e', 'Mixed_7c']
138
+ output : ['bottleneck_layer ', 'Mixed_5d', 'Mixed_6e', 'Mixed_7c']
139
139
wrapper : ' DistributedDataParallel'
140
140
requires_grad : True
141
141
optimizer :
@@ -161,7 +161,7 @@ train:
161
161
kwargs :
162
162
input :
163
163
is_from_teacher : False
164
- module_path : ' bottleneck '
164
+ module_path : ' bottleneck_layer '
165
165
io : ' output'
166
166
target :
167
167
is_from_teacher : True
0 commit comments