config: /root/in-context-learning-GLM/src/glm_configs/joint_training.yaml | |
model: | |
family: gpt2 | |
n_dims: 10 | |
n_embd: 256 | |
n_head: 8 | |
n_layer: 12 | |
n_positions: 40 | |
out_dir: gs://glm_weights/a648a15d-4896-4e95-8a8b-7150ff5ac620 | |
test_run: false | |
training: | |
batch_size: 256 | |
curriculum: | |
dims: | |
end: 10 | |
inc: 2 | |
interval: 2000 | |
start: 10 | |
points: | |
end: 40 | |
inc: 8 | |
interval: 2000 | |
start: 40 | |
data: gaussian | |
keep_every_steps: 10 | |
learning_rate: 0.00025 | |
num_tasks: 10000 | |
num_training_examples: 10000 | |
resume_id: null | |
save_every_steps: 10 | |
task: GLM | |
task_kwargs: | |
function_type: | |
- neg_binomial | |
- poisson | |
scaling: 0.32 | |
train_steps: 20000 | |
wandb: | |
entity: in-context | |
log_every_steps: 10 | |
name: null | |
notes: '' | |
project: in-context-training | |