File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change 1111from config_scripts .setup_config import read_config
1212from helper_scripts .ml_helpers import process_data , plot_confusion
1313from helper_scripts .ml_helpers import save_model
14-
14+ #TODO: Raise Not Implemented Error
15+ #TODO: This node is still in active development
1516
1617def _train_test_knn (df_processed : pd .DataFrame , sim_dict : dict , erlang : str ):
1718 predictor_df = df_processed ['num_segments' ]
Original file line number Diff line number Diff line change @@ -48,31 +48,39 @@ xt_noise = False
4848requested_xt = {" QPSK" : -26.19, " 16-QAM" : -36.69, " 64-QAM" : -41.69}
4949
5050[rl_settings]
51+ n_trials = 2
5152device = cpu
5253optimize = False
5354is_training = True
54- path_algorithm = ucb_bandit
55- path_model = greedy_bandit/NSFNet/0617/16_47_22_694727/state_vals_e750.0_routes_c4.json
55+
56+ path_algorithm = ppo
57+ path_model = None
5658core_algorithm = first_fit
57- core_model = greedy_bandit/NSFNet/0617/16_57_13_315030/state_vals_e750.0_cores_c4.json
59+ core_model = None
5860spectrum_algorithm = first_fit
59- spectrum_model = ppo/NSFNet/0512/12_57_55_484293
60- # Only for DRL
61+ spectrum_model = None
62+
6163render_mode = None
6264super_channel_space = 3
63- # Only for q-learning
64- learn_rate = 0.01
65- discount_factor = 0.95
66- epsilon_start = 0.2
67- epsilon_end = 0.05
68- reward = 1
69- penalty = -100
70- dynamic_reward = False
71- # TODO: Sim helpers has not been updated for this! (Only support for 2)
65+
66+ alpha_start = 0.1
67+ alpha_end = 0.01
68+ alpha_update = linear_decay
69+
70+ gamma = 0.1
71+
72+ epsilon_start = 0.0
73+ epsilon_end = 0.0
74+ epsilon_update = linear_decay
75+
7276path_levels = 2
7377decay_factor = 0.01
78+
79+ reward = 1
80+ penalty = -10
81+ dynamic_reward = False
7482core_beta = 0.1
75- gamma = 0.1
83+
7684
7785[ml_settings]
7886deploy_model = False
You can’t perform that action at this time.
0 commit comments