joint-training-r5.0 / config.yaml
connorchenn's picture
Upload folder using huggingface_hub
60bf39e verified
config: /root/in-context-learning-GLM/src/glm_configs/joint_training.yaml
model:
family: gpt2
n_dims: 10
n_embd: 256
n_head: 8
n_layer: 12
n_positions: 40
out_dir: gs://glm_weights/a648a15d-4896-4e95-8a8b-7150ff5ac620
test_run: false
training:
batch_size: 256
curriculum:
dims:
end: 10
inc: 2
interval: 2000
start: 10
points:
end: 40
inc: 8
interval: 2000
start: 40
data: gaussian
keep_every_steps: 10
learning_rate: 0.00025
num_tasks: 10000
num_training_examples: 10000
resume_id: null
save_every_steps: 10
task: GLM
task_kwargs:
function_type:
- neg_binomial
- poisson
scaling: 0.32
train_steps: 20000
wandb:
entity: in-context
log_every_steps: 10
name: null
notes: ''
project: in-context-training