Caleb Ellington commited on
Commit
fbd6f1d
·
1 Parent(s): 1622d28

update with major refactor

Browse files
fold0/config.yaml CHANGED
@@ -5,7 +5,7 @@ trainer:
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
- auto_wrap_policy: [genbio_finetune.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
@@ -118,10 +118,10 @@ trainer:
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
- class_path: genbio_finetune.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
- class_path: genbio_finetune.models.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
@@ -133,7 +133,7 @@ model:
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
- class_path: genbio_finetune.models.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
@@ -157,13 +157,13 @@ model:
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
- class_path: genbio_finetune.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
- class_path: genbio_finetune.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
 
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
+ auto_wrap_policy: [modelgenerator.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
 
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
+ class_path: modelgenerator.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
+ class_path: modelgenerator.backbones.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
 
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
+ class_path: modelgenerator.adapters.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
 
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
+ class_path: modelgenerator.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
+ class_path: modelgenerator.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
fold1/config.yaml CHANGED
@@ -5,7 +5,7 @@ trainer:
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
- auto_wrap_policy: [genbio_finetune.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
@@ -118,10 +118,10 @@ trainer:
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
- class_path: genbio_finetune.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
- class_path: genbio_finetune.models.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
@@ -133,7 +133,7 @@ model:
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
- class_path: genbio_finetune.models.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
@@ -157,13 +157,13 @@ model:
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
- class_path: genbio_finetune.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
- class_path: genbio_finetune.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
 
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
+ auto_wrap_policy: [modelgenerator.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
 
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
+ class_path: modelgenerator.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
+ class_path: modelgenerator.backbones.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
 
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
+ class_path: modelgenerator.adapters.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
 
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
+ class_path: modelgenerator.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
+ class_path: modelgenerator.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
fold2/config.yaml CHANGED
@@ -5,7 +5,7 @@ trainer:
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
- auto_wrap_policy: [genbio_finetune.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
@@ -118,10 +118,10 @@ trainer:
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
- class_path: genbio_finetune.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
- class_path: genbio_finetune.models.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
@@ -133,7 +133,7 @@ model:
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
- class_path: genbio_finetune.models.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
@@ -157,13 +157,13 @@ model:
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
- class_path: genbio_finetune.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
- class_path: genbio_finetune.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
 
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
+ auto_wrap_policy: [modelgenerator.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
 
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
+ class_path: modelgenerator.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
+ class_path: modelgenerator.backbones.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
 
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
+ class_path: modelgenerator.adapters.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
 
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
+ class_path: modelgenerator.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
+ class_path: modelgenerator.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
fold3/config.yaml CHANGED
@@ -5,7 +5,7 @@ trainer:
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
- auto_wrap_policy: [genbio_finetune.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
@@ -118,10 +118,10 @@ trainer:
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
- class_path: genbio_finetune.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
- class_path: genbio_finetune.models.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
@@ -133,7 +133,7 @@ model:
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
- class_path: genbio_finetune.models.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
@@ -157,13 +157,13 @@ model:
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
- class_path: genbio_finetune.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
- class_path: genbio_finetune.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
 
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
+ auto_wrap_policy: [modelgenerator.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
 
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
+ class_path: modelgenerator.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
+ class_path: modelgenerator.backbones.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
 
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
+ class_path: modelgenerator.adapters.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
 
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
+ class_path: modelgenerator.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
+ class_path: modelgenerator.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
fold4/config.yaml CHANGED
@@ -5,7 +5,7 @@ trainer:
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
- auto_wrap_policy: [genbio_finetune.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
@@ -118,10 +118,10 @@ trainer:
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
- class_path: genbio_finetune.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
- class_path: genbio_finetune.models.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
@@ -133,7 +133,7 @@ model:
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
- class_path: genbio_finetune.models.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
@@ -157,13 +157,13 @@ model:
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
- class_path: genbio_finetune.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
- class_path: genbio_finetune.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files:
 
5
  strategy:
6
  class_path: lightning.pytorch.strategies.FSDPStrategy
7
  init_args:
8
+ auto_wrap_policy: [modelgenerator.huggingface_models.fm4bio.modeling_fm4bio.FM4BioLayer]
9
  sharding_strategy: HYBRID_SHARD
10
  devices: 4
11
  num_nodes: 4
 
118
  reload_dataloaders_every_n_epochs: 0
119
  default_root_dir: logs
120
  model:
121
+ class_path: modelgenerator.tasks.SequenceRegression
122
  init_args:
123
  backbone:
124
+ class_path: modelgenerator.backbones.proteinfm
125
  init_args:
126
  from_scratch: false
127
  use_peft: true
 
133
  model_init_args: null
134
  max_length: 2048
135
  adapter:
136
+ class_path: modelgenerator.adapters.MLPPoolAdapter
137
  init_args:
138
  pooling: mean_pooling
139
  hidden_sizes:
 
157
  differentiable: false
158
  fused: null
159
  lr_scheduler:
160
+ class_path: modelgenerator.lr_schedulers.CosineWithWarmup
161
  init_args:
162
  warmup_ratio: 0.05
163
  strict_loading: true
164
  reset_optimizer_states: false
165
  data:
166
+ class_path: modelgenerator.data.DMSFitnessPrediction
167
  init_args:
168
  path: genbio-ai/ProteinGYM-DMS
169
  train_split_files: