Spaces:
Running
Running
Update src/plan.py
Browse files- src/plan.py +2 -7
src/plan.py
CHANGED
@@ -4,15 +4,10 @@ import json
|
|
4 |
import re
|
5 |
import os
|
6 |
|
7 |
-
model_id = "google/gemma-3n-E4B-it"
|
8 |
-
|
9 |
-
# Set Hugging Face cache directory
|
10 |
-
HF_CACHE_DIR = "./hf_cache"
|
11 |
-
os.environ["HF_HOME"] = HF_CACHE_DIR
|
12 |
-
os.environ["TRANSFORMERS_CACHE"] = HF_CACHE_DIR
|
13 |
-
os.makedirs(HF_CACHE_DIR, exist_ok=True)
|
14 |
|
15 |
hf_token = os.environ.get("HUGGINGFACE_TOKEN")
|
|
|
|
|
16 |
tokenizer = AutoTokenizer.from_pretrained(model_id, token=hf_token, use_auth_token=True)
|
17 |
model = AutoModelForCausalLM.from_pretrained(model_id, token=hf_token)
|
18 |
|
|
|
4 |
import re
|
5 |
import os
|
6 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
|
8 |
hf_token = os.environ.get("HUGGINGFACE_TOKEN")
|
9 |
+
model_id = "google/gemma-3n-E4B-it"
|
10 |
+
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_id, token=hf_token, use_auth_token=True)
|
12 |
model = AutoModelForCausalLM.from_pretrained(model_id, token=hf_token)
|
13 |
|