Skip to content

Commit bf458f6

Browse files
committed
adding submats
1 parent 68fdf92 commit bf458f6

55 files changed

Lines changed: 10676 additions & 3309 deletions

File tree

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

benchmark_configs/config_10_embeddings.yaml

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,9 @@ gradient_accumulation_steps: 2
1717

1818
use_muon: true
1919
muon_lr: 0.02 # Muon learning rate for hidden weights
20-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
20+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
21+
learning_rate: 0.0001
22+
2123

2224
seed: 42
2325
# Model Dimensions - 10 EMBEDDINGS
@@ -27,16 +29,16 @@ embedding_dim: 128
2729

2830

2931
hetero_gae: false # Use MultiMonoDecoder
30-
overwrite : true
32+
overwrite : false
3133

3234
# Encoder Configuration
3335
EMA: true
3436

3537
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3638
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
37-
use_commitment_scheduling: false # Enable scheduling
39+
use_commitment_scheduling: true # Enable scheduling
3840
commitment_schedule: linear # Smooth linear warmup
39-
commitment_warmup_steps: 10000 # Warmup over 1k steps (~5 epochs with batch_size=20)
41+
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4042
commitment_start: 0.05 # Start with very low commitment cost
4143

4244

benchmark_configs/config_15_embeddings.yaml

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ output_dir: "./models/embedding_comparison/final/"
66
run_name: "15_embeddings_all_decoders_final"
77

88

9-
overwrite : true
9+
overwrite : false
1010

1111
# Dataset
1212
dataset: "structs_train_final.h5"
@@ -20,7 +20,8 @@ gradient_accumulation_steps: 2
2020

2121
use_muon: true
2222
muon_lr: 0.02 # Muon learning rate for hidden weights
23-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
23+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
24+
learning_rate: 0.0001
2425

2526
seed: 42
2627

@@ -37,7 +38,7 @@ EMA: true
3738

3839
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3940
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
40-
use_commitment_scheduling: false # Enable scheduling
41+
use_commitment_scheduling: true # Enable scheduling
4142
commitment_schedule: linear # Smooth linear warmup
4243
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4344
commitment_start: 0.05 # Start with very low commitment cost

benchmark_configs/config_20_embeddings.yaml

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,8 @@ gradient_accumulation_steps: 2
1717

1818
use_muon: true
1919
muon_lr: 0.02 # Muon learning rate for hidden weights
20-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
20+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
21+
learning_rate: 0.0001
2122

2223
seed: 42
2324

@@ -28,14 +29,14 @@ embedding_dim: 128
2829

2930

3031
hetero_gae: false # Use MultiMonoDecoder
31-
overwrite : true
32+
overwrite : false
3233

3334
# Encoder Configuration
3435
EMA: true
3536

3637
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3738
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
38-
use_commitment_scheduling: false # Enable scheduling
39+
use_commitment_scheduling: true # Enable scheduling
3940
commitment_schedule: linear # Smooth linear warmup
4041
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4142
commitment_start: 0.05 # Start with very low commitment cost

benchmark_configs/config_25_embeddings.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -17,8 +17,8 @@ gradient_accumulation_steps: 2
1717

1818
use_muon: true
1919
muon_lr: 0.02 # Muon learning rate for hidden weights
20-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
21-
20+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
21+
learning_rate: 0.0001
2222

2323
seed: 42
2424

@@ -29,14 +29,14 @@ embedding_dim: 128
2929

3030

3131
hetero_gae: false # Use MultiMonoDecoder
32-
overwrite : true
32+
overwrite : false
3333

3434
# Encoder Configuration
3535
EMA: true
3636

3737
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3838
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
39-
use_commitment_scheduling: false # Enable scheduling
39+
use_commitment_scheduling: true # Enable scheduling
4040
commitment_schedule: linear # Smooth linear warmup
4141
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4242
commitment_start: 0.05 # Start with very low commitment cost

benchmark_configs/config_30_embeddings.yaml

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
# FoldTree2 Configuration: 30 Embeddings
22
# All three decoders: Sequence + Geometry + FoldX
33

4-
model_name: "ft2_30_embeddings_big"
4+
model_name: "ft2_30_embeddings"
55
output_dir: "./models/embedding_comparison/final/"
6-
run_name: "30_embeddings_all_decoders_big_final"
6+
run_name: "30_embeddings_all_decoders_final"
77

88
overwrite : false
99

@@ -19,8 +19,8 @@ gradient_accumulation_steps: 2
1919

2020
use_muon: true
2121
muon_lr: 0.02 # Muon learning rate for hidden weights
22-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
23-
learning_rate: 0.00001
22+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
23+
learning_rate: 0.0001
2424

2525
seed: 42
2626

@@ -31,7 +31,6 @@ embedding_dim: 128
3131

3232

3333
hetero_gae: false # Use MultiMonoDecoder
34-
overwrite : true
3534

3635
# Encoder Configuration
3736
EMA: true

benchmark_configs/config_35_embeddings.yaml

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
# FoldTree2 Configuration: 35 Embeddings
22
# All three decoders: Sequence + Geometry + FoldX
33

4-
model_name: "ft2_35_embeddings_big"
4+
model_name: "ft2_35_embeddings"
55
output_dir: "./models/embedding_comparison/final/"
6-
run_name: "35_embeddings_all_decoders_big_final"
6+
run_name: "35_embeddings_all_decoders_final"
77

88
overwrite : false
99

@@ -19,7 +19,8 @@ gradient_accumulation_steps: 2
1919

2020
use_muon: true
2121
muon_lr: 0.02 # Muon learning rate for hidden weights
22-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
22+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
23+
learning_rate: 0.0001
2324

2425
seed: 42
2526

@@ -30,16 +31,15 @@ embedding_dim: 128
3031

3132

3233
hetero_gae: false # Use MultiMonoDecoder
33-
overwrite : true
3434

3535
# Encoder Configuration
3636
EMA: true
3737

3838
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3939
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
40-
use_commitment_scheduling: false # Enable scheduling
40+
use_commitment_scheduling: true # Enable scheduling
4141
commitment_schedule: linear # Smooth linear warmup
42-
commitment_warmup_steps: 10000 # Warmup over 1k steps (~5 epochs with batch_size=20)
42+
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4343
commitment_start: 0.05 # Start with very low commitment cost
4444

4545

benchmark_configs/config_40_embeddings.yaml

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,9 @@ gradient_accumulation_steps: 2
1717

1818
use_muon: true
1919
muon_lr: 0.02 # Muon learning rate for hidden weights
20-
adamw_lr: 0.0001 # AdamW learning rate (1e-4 from notebook)
20+
adamw_lr: 0.0001 # AdamW learning rate (1e-5 from notebook)
21+
learning_rate: 0.0001
22+
2123

2224
seed: 42
2325

@@ -26,18 +28,17 @@ hidden_size: 150
2628
num_embeddings: 40
2729
embedding_dim: 128
2830

29-
3031
hetero_gae: false # Use MultiMonoDecoder
31-
overwrite : true
32+
overwrite : false
3233

3334
# Encoder Configuration
3435
EMA: true
3536

3637
# Commitment cost scheduling - THE FOCUS OF THIS CONFIG
3738
commitment_cost: 0.90 # Final commitment cost (high value for strong regularization)
38-
use_commitment_scheduling: false # Enable scheduling
39+
use_commitment_scheduling: true # Enable scheduling
3940
commitment_schedule: linear # Smooth linear warmup
40-
commitment_warmup_steps: 10000 # Warmup over 1k steps (~5 epochs with batch_size=20)
41+
commitment_warmup_steps: 1000 # Warmup over 1k steps (~5 epochs with batch_size=20)
4142
commitment_start: 0.05 # Start with very low commitment cost
4243

4344

config_notebook_1k_epochs.yaml

Lines changed: 15 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -7,21 +7,22 @@ dataset: "structs_train_final.h5"
77
output_dir: "./models/notebook_1k_training_nopositions_small/"
88
run_name: "notebook_replication_1k_epochs_small"
99

10-
overwrite: true
10+
overwrite: false
1111

1212
# Training hyperparameters (from notebook)
1313
epochs: 1000
14-
batch_size: 16
15-
gradient_accumulation_steps: 1
14+
batch_size: 8
15+
gradient_accumulation_steps: 2
1616
seed: 0
1717

1818
# Model architecture (from notebook)
19-
hidden_size: 150
19+
hidden_size: 200
2020
num_embeddings: 30
2121
embedding_dim: 128
2222

2323
# Encoder configuration (mk1_Encoder from notebook)
2424
EMA: true
25+
use_uncertainty_weighting: true
2526

2627
# Decoder configuration (MultiMonoDecoder from notebook)
2728
hetero_gae: false # Use MultiMonoDecoder
@@ -38,6 +39,9 @@ mixed_precision: true
3839
mask_plddt: true
3940
plddt_threshold: 0.3
4041

42+
jump_aa_loss: 25 # Ramp up amino acid loss weight after 25 epochs for stable training (from notebook)
43+
jump_ss_loss: 45 # Ramp up secondary structure loss weight after 45 epochs for stable training (from notebook)
44+
4145
# Learning rate scheduling (from notebook)
4246
lr_schedule: "plateau" # Notebook uses plateau scheduler
4347
lr_warmup_steps: 20
@@ -48,13 +52,15 @@ clip_grad: true
4852

4953
# Loss weights (from notebook cell)
5054
# Note: These are initial weights - notebook has weight schedulers (currently commented out)
51-
edgeweight: 0.1
52-
logitweight: 0.01
53-
xweight: 0.5
55+
edgeweight: .5
56+
logitweight: 0.1
57+
xweight: 0.1
5458
# fft2weight = 0.01
5559
vqweight: 0.1
56-
angles_weight: 0.1
57-
ss_weight: 0.1
60+
angles_weight: 0.01
61+
ss_weight: 0.01
62+
63+
5864

5965
# Commitment cost scheduling (from notebook encoder config)
6066
use_commitment_scheduling: true

0 commit comments

Comments
 (0)