JiakaiDu commited on
Commit
bd9c220
1 Parent(s): 1618811

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. Test_RAG.py +2 -2
Test_RAG.py CHANGED
@@ -60,7 +60,7 @@ if not text_example_cn_path.exists():
60
  f.write(content.read())
61
 
62
  model_language = "English"
63
- llm_model_id= "llama-3-8b-instruct"
64
  llm_model_configuration = SUPPORTED_LLM_MODELS[model_language][llm_model_id]
65
  print(f"Selected LLM model {llm_model_id}")
66
  prepare_int4_model = True # Prepare INT4 model
@@ -291,7 +291,7 @@ if llm_model_id == "red-pajama-3b-chat" and "GPU" in core.available_devices and
291
  ov_config["INFERENCE_PRECISION_HINT"] = "f32"
292
 
293
  llm = HuggingFacePipeline.from_model_id(
294
- model_id="meta-llama/Meta-Llama-3-8B",
295
  task="text-generation",
296
  backend="openvino",
297
  model_kwargs={
 
60
  f.write(content.read())
61
 
62
  model_language = "English"
63
+ llm_model_id= "llama-3.2-3b-instruct" #"llama-3-8b-instruct"
64
  llm_model_configuration = SUPPORTED_LLM_MODELS[model_language][llm_model_id]
65
  print(f"Selected LLM model {llm_model_id}")
66
  prepare_int4_model = True # Prepare INT4 model
 
291
  ov_config["INFERENCE_PRECISION_HINT"] = "f32"
292
 
293
  llm = HuggingFacePipeline.from_model_id(
294
+ model_id="meta-llama/Llama-3.2-3B-Instruct", #“meta-llama/Meta-Llama-3-8B"
295
  task="text-generation",
296
  backend="openvino",
297
  model_kwargs={