-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathtest_config.gin
49 lines (41 loc) · 1.45 KB
/
test_config.gin
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
from foo import bar as baz
include "blah"
# Parameters for AdamOptimizer:
# ==============================================================================
AdamOptimizer.beta1 = 0.9
AdamOptimizer.beta2 = 0.999
AdamOptimizer.epsilon = 0.001
AdamOptimizer.name = 'Adam'
AdamOptimizer.use_locking = False
AdamOptimizer.none_arg = None
# Parameters for exponential_decay:
# ==============================================================================
exponential_decay.learning_rate = 0.001
exponential_decay.decay_steps = 72926 # blah
exponential_decay.decay_rate = 0.92
# Parameters for NetworkTrainer:
# ==============================================================================
NetworkTrainer.dataset = @scope/module.test
NetworkTrainer.learning_rate = @exponential_decay, @linear_decay
NetworkTrainer.learning_rate_2 = [@exponential_decay, @linear_decay]
NetworkTrainer.foo = [ %bar, %baz ]
NetworkTrainer.network_fn = ( @inception_v3, @inception_v4 )
NetworkTrainer.optimizer = @AdamOptimizer
NetworkTrainer.train_steps = 1000000
multiline = [
%bar,
@test,
]
my_dict_singleline = { blah: 0, bloh: 12, @random: %whatelse}
my_dict = {
blah: 0,
bloh: 12,
@random: [%whatelse, %whatelse],
@random2: %anything,
"bladder": ['b473f254"-81cf-4a41-95d4-f2d9a476b6b6'],
}
num_layers/blah/macro.value = 10
network.num_layers = @num_layers/macro()
num_layers_2 = 10
network.num_layers = %num_layers_2
network.num_layers_with_scope = %scope.foo.bar