fix typo
Browse files
README.md
CHANGED
@@ -371,7 +371,7 @@ The accuracy is evaluated on CUDA with overflow protection, and it is expected t
|
|
371 |
~~~python
|
372 |
import transformers
|
373 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
374 |
-
|
375 |
|
376 |
# https://github.com/huggingface/transformers/pull/35493
|
377 |
def set_initialized_submodules(model, state_dict_keys):
|
@@ -419,7 +419,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
419 |
trust_remote_code=True,
|
420 |
device_map=device_map,
|
421 |
)
|
422 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
423 |
|
424 |
|
425 |
def forward_hook(module, input, output):
|
|
|
371 |
~~~python
|
372 |
import transformers
|
373 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
374 |
+
from lm_eval.utils import make_table
|
375 |
|
376 |
# https://github.com/huggingface/transformers/pull/35493
|
377 |
def set_initialized_submodules(model, state_dict_keys):
|
|
|
419 |
trust_remote_code=True,
|
420 |
device_map=device_map,
|
421 |
)
|
422 |
+
tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
|
423 |
|
424 |
|
425 |
def forward_hook(module, input, output):
|