Skip to content

Commit 39a6519

Browse files
committed
Fix:MINOR
1 parent 47691e3 commit 39a6519

11 files changed

Lines changed: 154 additions & 145 deletions

File tree

config/CNN/NP/README

Whitespace-only changes.

config/CNN/NP/data_spec.json

Lines changed: 10 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,33 +1,36 @@
11
{
22
"validation": {
3-
"base_path" : "data/NPFILE",
3+
"base_path" : "/data/abil/MNIST/NP",
44
"filename" : "val.dat",
5-
"partition" : 200,
5+
"partition" : 100,
66
"random" : true,
77
"random_seed" : 123,
88
"dim_shuffle" : [0,3,2,1],
9+
"batch_size": 512,
910
"keep_flatten" : false,
1011
"reader_type" : "NP"
1112
},
1213

1314
"training" : {
14-
"base_path" : "data/NPFILE",
15+
"base_path" : "/data/abil/MNIST/NP",
1516
"filename" : "train.dat",
16-
"partition" : 200,
17+
"partition" : 100,
1718
"random" : true,
1819
"random_seed" : 123,
1920
"dim_shuffle" : [0,3,2,1],
21+
"batch_size": 512,
2022
"keep_flatten" : false,
2123
"reader_type" : "NP"
2224
},
2325

2426
"testing" : {
25-
"base_path" : "data/NPFILE",
26-
"filename" : "train.dat",
27-
"partition" : 200,
27+
"base_path" : "/data/abil/MNIST/NP",
28+
"filename" : "test.dat",
29+
"partition" : 100,
2830
"random" : true,
2931
"random_seed" : 123,
3032
"dim_shuffle" : [0,3,2,1],
33+
"batch_size": 512,
3134
"keep_flatten" : false,
3235
"reader_type" : "NP"
3336
}

config/CNN/NP/model_conf.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
"output_file" : "conv_out.model",
1919

2020
"comment" : "input_shape :: (Mandatory) specify the input shape of a given feature vector" ,
21-
"input_shape" : [3,28,28],
21+
"input_shape" : [1,28,28],
2222

2323
"comment" : "batch_size :: specify the mini batch size while training, default 256",
2424
"batch_size" : 256,
@@ -27,7 +27,7 @@
2727
"momentum" : 0.5,
2828

2929
"comment" : "n_outs :: Specify the number of outputs",
30-
"n_outs" : 8,
30+
"n_outs" : 10,
3131

3232
"comment" : "l_rate_method :: Two method l_rate method are supported C: Constant learning rate and E : Exponential decay",
3333
"l_rate_method" : "C",

config/CNN/NP/nnet_spec.json

Lines changed: 30 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -1,37 +1,37 @@
11
{
2-
"comment" : "layers :: convolution layer configuration",
3-
"cnn": {
4-
"layers": [
5-
{
6-
"convmat_dim" : [5,5],
7-
"num_filters" : 100,
8-
"poolsize" : [2,2],
9-
"flatten" : false,
10-
"update" : true
11-
},
12-
{
13-
"convmat_dim" : [4,4],
14-
"num_filters" : 200,
15-
"poolsize" : [2,2],
16-
"flatten" : true,
17-
"update" : true
18-
}
19-
],
2+
"comment" : "layers :: convolution layer configuration",
3+
"cnn": {
4+
"layers": [
5+
{
6+
"convmat_dim" : [5,5],
7+
"num_filters" : 100,
8+
"poolsize" : [2,2],
9+
"flatten" : false,
10+
"update" : true
11+
},
12+
{
13+
"convmat_dim" : [4,4],
14+
"num_filters" : 200,
15+
"poolsize" : [2,2],
16+
"flatten" : true,
17+
"update" : true
18+
}
19+
],
2020

21-
"comment" : "activation :: sigmoid or tanh",
22-
"activation" : "tanh",
21+
"comment" : "activation :: sigmoid or tanh",
22+
"activation" : "tanh",
2323

24-
"comment" : "use_fast :: To use pylearn2 library for faster computation or not",
25-
"use_fast" : false
26-
},
24+
"comment" : "use_fast :: To use pylearn2 library for faster computation or not",
25+
"use_fast" : false
26+
},
27+
28+
"comment" : "layers :: hidden layer configuration",
29+
"mlp" : {
2730

28-
"comment" : "layers :: hidden layer configuration",
29-
"mlp" : {
30-
31-
"layers": [500,200],
31+
"layers": [500,200],
3232

33-
"comment" : "activation :: sigmoid or tanh",
34-
"activation" : "tanh"
35-
}
33+
"comment" : "activation :: sigmoid or tanh",
34+
"activation" : "tanh"
35+
}
3636
}
3737

config/DBN/data_spec.json

Lines changed: 30 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -1,31 +1,34 @@
11
{
2-
"validation": {
3-
"base_path" : "data/NPFILE",
4-
"filename" : "val.dat",
5-
"partition" : 200,
6-
"random" : true,
7-
"random_seed" : 123,
8-
"keep_flatten" : true,
9-
"reader_type" : "NP"
10-
},
2+
"validation": {
3+
"base_path" : "/data/abil/MNIST/NP",
4+
"filename" : "val.dat",
5+
"partition" : 100,
6+
"random" : true,
7+
"random_seed" : 123,
8+
"batch_size": 512,
9+
"keep_flatten" : true,
10+
"reader_type" : "NP"
11+
},
1112

12-
"training" : {
13-
"base_path" : "data/NPFILE",
14-
"filename" : "train.dat",
15-
"partition" : 200,
16-
"random" : true,
17-
"random_seed" : 123,
18-
"keep_flatten" : true,
19-
"reader_type" : "NP"
20-
},
13+
"training" : {
14+
"base_path" : "/data/abil/MNIST/NP",
15+
"filename" : "train.dat",
16+
"partition" : 100,
17+
"random" : true,
18+
"random_seed" : 123,
19+
"batch_size": 512,
20+
"keep_flatten" : true,
21+
"reader_type" : "NP"
22+
},
2123

22-
"testing" : {
23-
"base_path" : "data/NPFILE",
24-
"filename" : "train.dat",
25-
"partition" : 200,
26-
"random" : true,
27-
"random_seed" : 123,
28-
"keep_flatten" : true,
29-
"reader_type" : "NP"
30-
}
24+
"testing" : {
25+
"base_path" : "/data/abil/MNIST/NP",
26+
"filename" : "test.dat",
27+
"partition" : 100,
28+
"random" : true,
29+
"random_seed" : 123,
30+
"batch_size": 512,
31+
"keep_flatten" : true,
32+
"reader_type" : "NP"
33+
}
3134
}

config/DBN/model_conf.json

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -18,10 +18,10 @@
1818
"batch_size" : 128,
1919

2020
"comment": "",
21-
"n_ins":2352,
21+
"n_ins":784,
2222

2323
"comment":"",
24-
"n_outs":200,
24+
"n_outs":10,
2525

2626
"comment" :"TODO",
2727
"gbrbm_learning_rate":0.005,
@@ -53,8 +53,8 @@
5353
"finetune_momentum": 0.5,
5454

5555
"processes":{
56-
"pretraining":false,
57-
"finetuning":false,
56+
"pretraining":true,
57+
"finetuning":true,
5858
"testing":true,
5959
"export_data":false
6060
}

config/DBN/rbm_spec.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,13 +2,13 @@
22
{
33

44
"comment" : "layers :: RBM layer configuration (No: of Nodes)",
5-
"hidden_layers": [ 2350,1024,1024,1024,1024,1901 ],
5+
"hidden_layers": [1000, 1000, 1000],
66

77
"comment" : "activation :: sigmoid or tanh",
88
"activation" : "sigmoid",
99

1010
"comment" : "pretrained_layers:number of layers to be pre-trained",
11-
"pretrained_layers" : 5,
11+
"pretrained_layers" : 3,
1212

1313
"comment" : "first_layer_type::type for the first layer; either 'bb' (Bernoulli-Bernoulli) or 'gb' (Gaussian-Bernoulli)",
1414
"first_layer_type" : "gb",

config/SDA/data_spec.json

Lines changed: 10 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,30 +1,33 @@
11
{
22
"validation": {
3-
"base_path" : "data/NPFILE",
3+
"base_path" : "/data/abil/MNIST/NP",
44
"filename" : "val.dat",
5-
"partition" : 200,
5+
"partition" : 100,
66
"random" : true,
77
"random_seed" : 123,
8+
"batch_size": 512,
89
"keep_flatten" : true,
910
"reader_type" : "NP"
1011
},
1112

1213
"training" : {
13-
"base_path" : "data/NPFILE",
14+
"base_path" : "/data/abil/MNIST/NP",
1415
"filename" : "train.dat",
15-
"partition" : 200,
16+
"partition" : 100,
1617
"random" : true,
1718
"random_seed" : 123,
19+
"batch_size": 512,
1820
"keep_flatten" : true,
1921
"reader_type" : "NP"
2022
},
2123

2224
"testing" : {
23-
"base_path" : "data/NPFILE",
24-
"filename" : "train.dat",
25-
"partition" : 200,
25+
"base_path" : "/data/abil/MNIST/NP",
26+
"filename" : "test.dat",
27+
"partition" : 100,
2628
"random" : true,
2729
"random_seed" : 123,
30+
"batch_size": 512,
2831
"keep_flatten" : true,
2932
"reader_type" : "NP"
3033
}

config/SDA/model_conf.json

Lines changed: 57 additions & 56 deletions
Original file line numberDiff line numberDiff line change
@@ -1,58 +1,59 @@
11
{
2-
"comment" : "nnetType :: (Mandatory) specify Type of Network (CNN,SDA) ",
3-
"nnetType" : "SDA",
4-
5-
"comment" : "train_data :: (Mandatory) specify the working directory containing data configuration and output ",
6-
"wdir" : "wdir",
7-
8-
"comment" : "valid_data (Mandatory) specify the path of the validation data relative to the working directory",
9-
"data_spec" : "data_spec.json",
10-
11-
"comment" : "sda_nnet_spec:: (Mandatory) specify the path of SDA network configuration specification relative to working directory",
12-
"nnet_spec" : "sda_spec.json",
13-
14-
"comment" : "output_file :: (Mandatory) specify the path of SDA network output file relative to working directory",
15-
"output_file" : "sda_out.model",
16-
17-
"comment" : "batch_size :: specify the mini batch size while training, default 128",
18-
"batch_size" : 128,
19-
20-
"comment": "",
21-
"n_ins":2352,
22-
23-
"comment":"",
24-
"n_outs":200,
25-
26-
"comment" :"pretraining_epochs: number of epoch to do pretraining",
27-
"pretraining_epochs":10,
28-
29-
30-
"comment" :"learning rate to be used during pre-training",
31-
"pretrain_lr":0.08,
32-
33-
"comment" : "finetune_method:: Two methods are supported C: Constant learning rate and E : Exponential decay",
34-
"finetune_method":"C",
35-
36-
"comment" : "finetune_rate :: learning rate configuration",
37-
"finetune_rate" : {
38-
"learning_rate" : 0.08,
39-
"epoch_num" : 10,
40-
41-
"start_rate" : 0.08,
42-
"scale_by" : 0.5,
43-
"min_derror_decay_start" : 0.05,
44-
"min_derror_stop" : 0.05,
45-
"min_epoch_decay_start" : 15,
46-
"init_error" :100
47-
},
48-
49-
"comment" : "finetune_momentum :: Specify the momentum factor while finetuning",
50-
"finetune_momentum": 0.5,
51-
52-
"processes":{
53-
"pretraining":true,
54-
"finetuning":true,
55-
"testing":true,
56-
"export_data":false
57-
}
2+
"comment" : "nnetType :: (Mandatory) specify Type of Network (CNN,SDA) ",
3+
"nnetType" : "SDA",
4+
5+
"comment" : "train_data :: (Mandatory) specify the working directory containing data configuration and output ",
6+
"wdir" : "wdir",
7+
8+
"comment" : "valid_data (Mandatory) specify the path of the validation data relative to the working directory",
9+
"data_spec" : "data_spec.json",
10+
11+
"comment" : "sda_nnet_spec:: (Mandatory) specify the path of SDA network configuration specification relative to working directory",
12+
"nnet_spec" : "sda_spec.json",
13+
14+
"comment" : "output_file :: (Mandatory) specify the path of SDA network output file relative to working directory",
15+
"output_file" : "sda_out.model",
16+
17+
"logger_level":"INFO",
18+
19+
"comment" : "batch_size :: specify the mini batch size while training, default 128",
20+
"batch_size" : 128,
21+
22+
"comment": "",
23+
"n_ins":784,
24+
25+
"comment":"",
26+
"n_outs":10,
27+
28+
"comment" :"pretraining_epochs: number of epoch to do pretraining",
29+
"pretraining_epochs":10,
30+
31+
"comment" :"learning rate to be used during pre-training",
32+
"pretrain_lr":0.08,
33+
34+
"comment" : "finetune_method:: Two methods are supported C: Constant learning rate and E : Exponential decay",
35+
"finetune_method":"C",
36+
37+
"comment" : "finetune_rate :: learning rate configuration",
38+
"finetune_rate" : {
39+
"learning_rate" : 0.08,
40+
"epoch_num" : 10,
41+
42+
"start_rate" : 0.08,
43+
"scale_by" : 0.5,
44+
"min_derror_decay_start" : 0.05,
45+
"min_derror_stop" : 0.05,
46+
"min_epoch_decay_start" : 15,
47+
"init_error" :100
48+
},
49+
50+
"comment" : "finetune_momentum :: Specify the momentum factor while finetuning",
51+
"finetune_momentum": 0.5,
52+
53+
"processes":{
54+
"pretraining":true,
55+
"finetuning":true,
56+
"testing":true,
57+
"export_data":false
58+
}
5859
}

0 commit comments

Comments
 (0)