from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("Isotonic/gpt-human-assistant") def tokenize_function(examples): return tokenizer(examples["prompt"], truncation=True) tokenized_datasets = dataset.map(tokenize_function, batched=True)