gowthamtamilkathir commited on
Commit
00ff3dc
·
verified ·
1 Parent(s): c3f9463

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -3
app.py CHANGED
@@ -1,20 +1,27 @@
1
  from fastapi import FastAPI, Query
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
 
4
 
5
  app = FastAPI()
6
 
7
- # Load the model and tokenizer
 
 
 
8
  tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-llm-7b-base")
9
  model = AutoModelForCausalLM.from_pretrained(
10
  "deepseek-ai/deepseek-llm-7b-base",
11
  torch_dtype=torch.float16,
12
- device_map="auto"
 
13
  )
14
 
15
  @app.get("/")
16
  def home():
17
- return {"message": "DeepSeek LLM is running. Use /ask?prompt=your+question"}
 
 
18
 
19
  @app.get("/ask")
20
  def ask(prompt: str = Query(..., description="Your input prompt")):
 
1
  from fastapi import FastAPI, Query
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
4
+ import os
5
 
6
  app = FastAPI()
7
 
8
+ # Create offload folder if not exists
9
+ os.makedirs("./offload", exist_ok=True)
10
+
11
+ # Load tokenizer and model with offload_folder to prevent device_map error
12
  tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-llm-7b-base")
13
  model = AutoModelForCausalLM.from_pretrained(
14
  "deepseek-ai/deepseek-llm-7b-base",
15
  torch_dtype=torch.float16,
16
+ device_map="auto",
17
+ offload_folder="./offload"
18
  )
19
 
20
  @app.get("/")
21
  def home():
22
+ return {
23
+ "message": "✅ DeepSeek LLM is running. Use endpoint /ask?prompt=your+question"
24
+ }
25
 
26
  @app.get("/ask")
27
  def ask(prompt: str = Query(..., description="Your input prompt")):