Planted in Pretraining
Collection
Models and dataset from the CoLM 2025 paper : "Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs
•
14 items
•
Updated
Model Description
This 🤗 Transformers model was finetuned using LoRA adapters for the arXiv paper:
"Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs"
We study whether cognitive biases in LLMs emerge from pretraining, instruction tuning, or training randomness.
This is one of 3 identical versions trained with different random seeds.
allenai/OLMo-7B
For research on cognitive biases in LLMs. Used to test causal impact of pretraining vs instruction tuning.
Do not use in production, sensitive domains, or decision-critical applications.
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("itay1itzhak/OLMo-Tulu-Seed-0")
tokenizer = AutoTokenizer.from_pretrained("itay1itzhak/OLMo-Tulu-Seed-0")
inputs = tokenizer("Example input?", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0]))
Base model
allenai/OLMo-7B