cicdatopea commited on
Commit
9b8cc6e
·
verified ·
1 Parent(s): cfbb6b2
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -371,7 +371,7 @@ The accuracy is evaluated on CUDA with overflow protection, and it is expected t
371
  ~~~python
372
  import transformers
373
  from transformers import AutoModelForCausalLM, AutoTokenizer
374
-
375
 
376
  # https://github.com/huggingface/transformers/pull/35493
377
  def set_initialized_submodules(model, state_dict_keys):
@@ -419,7 +419,7 @@ model = AutoModelForCausalLM.from_pretrained(
419
  trust_remote_code=True,
420
  device_map=device_map,
421
  )
422
- tokenizer = AutoTokenizer.from_pretrained(model_name)
423
 
424
 
425
  def forward_hook(module, input, output):
 
371
  ~~~python
372
  import transformers
373
  from transformers import AutoModelForCausalLM, AutoTokenizer
374
+ from lm_eval.utils import make_table
375
 
376
  # https://github.com/huggingface/transformers/pull/35493
377
  def set_initialized_submodules(model, state_dict_keys):
 
419
  trust_remote_code=True,
420
  device_map=device_map,
421
  )
422
+ tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
423
 
424
 
425
  def forward_hook(module, input, output):