Skip to content

Commit c322135

Browse files
committed
Fix:Minor:FILE
1 parent dfa022e commit c322135

8 files changed

Lines changed: 80 additions & 161 deletions

File tree

config/DBN/data_spec.json

Lines changed: 24 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,19 +1,37 @@
11
{
22
"validation": {
3-
"path" : "data/val.dat",
3+
"base_path" : "data/NPFILE",
4+
"filename" : "val.dat",
45
"partition" : 200,
56
"random" : true,
67
"random_seed" : 123,
7-
"keep_flatten": true,
8-
"batch_size": 256
8+
"dim_shuffle" : [0,3,2,1],
9+
"batch_size": 256,
10+
"keep_flatten" : true,
11+
"reader_type" : "NP"
912
},
1013

1114
"training" : {
12-
"path" : "data/train.dat",
15+
"base_path" : "data/NPFILE",
16+
"filename" : "train.dat",
1317
"partition" : 200,
1418
"random" : true,
1519
"random_seed" : 123,
16-
"keep_flatten": true,
17-
"batch_size": 256
20+
"dim_shuffle" : [0,3,2,1],
21+
"batch_size": 256,
22+
"keep_flatten" : true,
23+
"reader_type" : "NP"
24+
},
25+
26+
"testing" : {
27+
"base_path" : "data/NPFILE",
28+
"filename" : "train.dat",
29+
"partition" : 200,
30+
"random" : true,
31+
"random_seed" : 123,
32+
"dim_shuffle" : [0,3,2,1],
33+
"batch_size": 256,
34+
"keep_flatten" : true,
35+
"reader_type" : "NP"
1836
}
1937
}

config/DBN/mlp_spec.json

Lines changed: 0 additions & 8 deletions
This file was deleted.

config/SDA/data_spec.json

Lines changed: 24 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,19 +1,37 @@
11
{
22
"validation": {
3-
"path" : "data/val.dat",
3+
"base_path" : "data/NPFILE",
4+
"filename" : "val.dat",
45
"partition" : 200,
56
"random" : true,
67
"random_seed" : 123,
7-
"keep_flatten": true,
8-
"batch_size": 256
8+
"dim_shuffle" : [0,3,2,1],
9+
"batch_size": 256,
10+
"keep_flatten" : true,
11+
"reader_type" : "NP"
912
},
1013

1114
"training" : {
12-
"path" : "data/train.dat",
15+
"base_path" : "data/NPFILE",
16+
"filename" : "train.dat",
1317
"partition" : 200,
1418
"random" : true,
1519
"random_seed" : 123,
16-
"keep_flatten": true,
17-
"batch_size": 256
20+
"dim_shuffle" : [0,3,2,1],
21+
"batch_size": 256,
22+
"keep_flatten" : true,
23+
"reader_type" : "NP"
24+
},
25+
26+
"testing" : {
27+
"base_path" : "data/NPFILE",
28+
"filename" : "train.dat",
29+
"partition" : 200,
30+
"random" : true,
31+
"random_seed" : 123,
32+
"dim_shuffle" : [0,3,2,1],
33+
"batch_size": 256,
34+
"keep_flatten" : true,
35+
"reader_type" : "NP"
1836
}
1937
}

io_modules/__init__.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,10 +23,12 @@ def setLogger(level="INFO",stderr=True,name=None,logFile='python-dnn.log'):
2323
# Set the level which determines what you see
2424
if level == "INFO":
2525
logger.setLevel(logging.INFO)
26-
elif level=="DEBUG":
26+
elif level == "DEBUG":
2727
logger.setLevel(logging.DEBUG)
28+
elif level == "ERROR":
29+
logger.setLevel(logging.ERROR)
2830
else:
29-
logger.setLevel(logging.ERROR)
31+
logger.setLevel(logging.WARNING)
3032

3133
def create_folder_structure_if_not_exists(filepath):
3234
import logging,os,errno

io_modules/file_io.py

Lines changed: 0 additions & 129 deletions
This file was deleted.

io_modules/file_reader.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -165,6 +165,7 @@ def read_next_partition_data(self,already_read=0,pad_zeros=False):
165165
self.partition_num = self.partition_num + 1
166166
logger.debug('TD Filereader : %d frames read from %s' % (self.cur_frame_num,self.filepath))
167167
if not self.options['keep_flatten'] : #reshape the vector if needed
168+
logger.debug('TD Filereader : Reshape input...')
168169
shape = [self.cur_frame_num];
169170
shape.extend(self.options['input_shape']);
170171
self.feat = self.feat.reshape(shape);

run_DBN.py

Lines changed: 16 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -26,9 +26,10 @@
2626
#module imports
2727
from utils.load_conf import load_model,load_rbm_spec,load_data_spec
2828
from models.dbn import DBN
29-
from io_modules.file_io import read_dataset
29+
from io_modules.file_reader import read_dataset
3030
from io_modules import setLogger
3131
from utils.learn_rates import LearningRate
32+
from io_modules.model_io import _nnet2file, _file2nnet
3233

3334
from models import fineTunning,testing
3435

@@ -73,7 +74,8 @@ def preTraining(nnetModel,train_sets,train_xy,train_x,train_y,model_config):
7374
while not train_sets.is_finish():
7475
train_sets.make_partition_shared(train_xy)
7576
#train_sets.load_next_partition(train_xy)
76-
for batch_index in xrange(train_sets.cur_frame_num / batch_size): # loop over mini-batches
77+
for batch_index in xrange(train_sets.cur_frame_num / batch_size):
78+
# loop over mini-batches
7779
#logger.info("Training For epoch %d and batch %d",epoch,batch_index)
7880
[reconstruction_cost, free_energy_cost] = pretraining_fns[i](index=batch_index,
7981
lr=pretrain_lr,
@@ -82,7 +84,8 @@ def preTraining(nnetModel,train_sets,train_xy,train_x,train_y,model_config):
8284
fe_c.append(free_energy_cost)
8385
train_sets.read_next_partition_data()
8486
train_sets.initialize_read()
85-
logger.info('Training layer %i, epoch %d, r_cost %f, fe_cost %f' % (i, epoch, numpy.mean(r_c), numpy.mean(fe_c)))
87+
logger.info('Training layer %i, epoch %d, r_cost %f, fe_cost %f',
88+
i, epoch, numpy.mean(r_c), numpy.mean(fe_c))
8689
end_time = time.clock()
8790
logger.info('The PreTraing ran for %.2fm' % ((end_time - start_time) / 60.))
8891

@@ -105,19 +108,22 @@ def runRBM(configFile):
105108

106109
train_sets, train_xy, train_x, train_y = read_dataset(data_spec['training'])
107110

108-
keep_layer_num=model_config['keep_layer_num']
109-
batch_size=model_config['batch_size']
111+
keep_layer_num = model_config['keep_layer_num']
112+
batch_size = model_config['batch_size']
113+
wdir = model_config['wdir']
110114

111115
if keep_layer_num > 0:
112-
#current_nnet = wdir + 'nnet.ptr.current'
116+
current_nnet = wdir + '/nnet.ptr.current'
113117
logger.info('Initializing model from ' + str(current_nnet) + '....')
114118
# load model
115-
#_file2nnet(dbn.sigmoid_layers, set_layer_num = keep_layer_num, filename = current_nnet, withfinal=False)
119+
_file2nnet(dbn.sigmoid_layers, set_layer_num = keep_layer_num,
120+
filename = current_nnet, withfinal=False)
116121

117122
preTraining(dbn,train_sets,train_xy,train_x,train_y,model_config)
118123

119124
# save the pretrained nnet to file
120-
#_nnet2file(dbn.sigmoid_layers, filename=output_file, withfinal=True)
125+
logger.info('Saving model to ' + str(model_config['output_file']) + '....')
126+
_nnet2file(dbn.sigmoid_layers, filename=model_config['output_file'], withfinal=True)
121127

122128

123129
########################
@@ -146,6 +152,8 @@ def runRBM(configFile):
146152
fineTunning(dbn,train_sets,train_xy,train_x,train_y,
147153
valid_sets,valid_xy,valid_x,valid_y,lrate,momentum,batch_size)
148154

155+
logger.info('Saving model to ' + str(model_config['output_file']) + '.final ....')
156+
_nnet2file(dbn.sigmoid_layers, filename=model_config['output_file']+'.final', withfinal=True)
149157

150158
try:
151159
test_sets, test_xy, test_x, test_y = read_dataset(data_spec['testing'])

run_SDA.py

Lines changed: 11 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,11 +23,11 @@
2323

2424
#module imports
2525
from utils.load_conf import load_model,load_sda_spec,load_data_spec
26-
from io_modules.file_io import read_dataset
26+
from io_modules.file_reader import read_dataset
2727
from io_modules import setLogger
2828
from utils.learn_rates import LearningRate
2929
from models.sda import SDA
30-
30+
from io_modules.model_io import _nnet2file, _file2nnet
3131
from models import fineTunning,testing
3232

3333

@@ -104,6 +104,10 @@ def runSdA(configFile):
104104

105105
logger.info('The PreTraing ran for %.2fm' % ((end_time - start_time) / 60.))
106106

107+
# save the pretrained nnet to file
108+
logger.info('Saving model to ' + str(model_config['output_file']) + '....')
109+
_nnet2file(dbn.sigmoid_layers, filename=model_config['output_file'], withfinal=True)
110+
107111

108112
########################
109113
# FINETUNING THE MODEL #
@@ -130,6 +134,11 @@ def runSdA(configFile):
130134
fineTunning(sda,train_sets,train_xy,train_x,train_y,
131135
valid_sets,valid_xy,valid_x,valid_y,lrate,momentum,batch_size);
132136

137+
# save the pretrained nnet to file
138+
logger.info('Saving model to ' + str(model_config['output_file']) + '.final ....')
139+
_nnet2file(dbn.sigmoid_layers, filename=model_config['output_file']+'.final', withfinal=True)
140+
141+
133142
try:
134143
test_sets, test_xy, test_x, test_y = read_dataset(data_spec['testing'])
135144
except KeyError:

0 commit comments

Comments
 (0)