Planted in Pretraining
Collection
Models and dataset from the CoLM 2025 paper : "Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs
•
14 items
•
Updated
Model Description
This 🤗 Transformers model was finetuned using LoRA adapters for the arXiv paper:
"Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs"
We study whether cognitive biases in LLMs emerge from pretraining, instruction tuning, or training randomness.
This is one of 3 identical versions trained with different random seeds.
google/t5-v1_1-xxl
For research on cognitive biases in LLMs. Used to test causal impact of pretraining vs instruction tuning.
Do not use in production, sensitive domains, or decision-critical applications.
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("itay1itzhak/T5-Tulu-Seed-2")
tokenizer = AutoTokenizer.from_pretrained("itay1itzhak/T5-Tulu-Seed-2")
inputs = tokenizer("Example input?", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0]))
@misc{itzhak2025plantedpretrainingswayedfinetuning,
title={Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs},
author={Itay Itzhak and Yonatan Belinkov and Gabriel Stanovsky},
year={2025},
eprint={2507.07186},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2507.07186},
}