Files changed (1) hide show
  1. README.md +24 -1
README.md CHANGED
@@ -1,3 +1,26 @@
1
  ---
2
  license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ ---
4
+
5
+
6
+ To run a simple model do the following. The model of course didnt work that well for me:
7
+ pip install llama-cpp-python
8
+ pip install huggingface-hub
9
+
10
+ You could of course change the gguf file to download. Please dont download all the files as it can be fairly big.
11
+ huggingface-cli download aaditya/OpenBioLLM-Llama3-8B-GGUF openbiollm-llama3-8b.Q4_K_M.gguf --local-dir ./models --local-dir-use-symlinks False
12
+
13
+ The file to simply start generating, do the following:
14
+
15
+ from llama_cpp import Llama
16
+ llm = Llama(model_path="./models/openbiollm-llama3-8b.Q4_K_M.gguf", chat_format="llama-3") # Set chat_format according to the model you are using
17
+ response=llm.create_chat_completion(
18
+ max_tokens=250, messages = [
19
+ {"role": "system", "content": "You are biomed ai"},
20
+ {"role": "user", "content": "name 5 diabetes medications"}
21
+
22
+ ]
23
+ )
24
+
25
+ print(response["choices"][0]["message"]["content"])
26
+