Spaces:
Sleeping
Sleeping
numpy issue debugging
Browse files- Dockerfile +0 -1
- app.py +2 -0
- qwen_classifier/evaluate.py +1 -1
- requirements.txt +1 -0
Dockerfile
CHANGED
|
@@ -22,7 +22,6 @@ COPY app.py .
|
|
| 22 |
COPY qwen_classifier/ ./qwen_classifier/
|
| 23 |
COPY setup.py .
|
| 24 |
RUN pip install --no-cache-dir -r requirements.txt
|
| 25 |
-
RUN pip install --no-cache-dir numpy
|
| 26 |
|
| 27 |
# Install PyTorch with CUDA support
|
| 28 |
RUN pip install --no-cache-dir \
|
|
|
|
| 22 |
COPY qwen_classifier/ ./qwen_classifier/
|
| 23 |
COPY setup.py .
|
| 24 |
RUN pip install --no-cache-dir -r requirements.txt
|
|
|
|
| 25 |
|
| 26 |
# Install PyTorch with CUDA support
|
| 27 |
RUN pip install --no-cache-dir \
|
app.py
CHANGED
|
@@ -6,6 +6,7 @@ os.makedirs(os.environ['HF_HOME'], exist_ok=True) # Ensure directory exists
|
|
| 6 |
from fastapi import FastAPI
|
| 7 |
from fastapi.responses import HTMLResponse
|
| 8 |
import torch
|
|
|
|
| 9 |
from transformers import AutoTokenizer
|
| 10 |
from huggingface_hub import login
|
| 11 |
from pydantic import BaseModel
|
|
@@ -18,6 +19,7 @@ from qwen_classifier.globals import global_model, global_tokenizer
|
|
| 18 |
from qwen_classifier.model import QwenClassifier
|
| 19 |
from qwen_classifier.config import HF_REPO, DEVICE
|
| 20 |
|
|
|
|
| 21 |
|
| 22 |
app = FastAPI(title="Qwen Classifier")
|
| 23 |
hf_repo = os.getenv("HF_REPO")
|
|
|
|
| 6 |
from fastapi import FastAPI
|
| 7 |
from fastapi.responses import HTMLResponse
|
| 8 |
import torch
|
| 9 |
+
import numpy
|
| 10 |
from transformers import AutoTokenizer
|
| 11 |
from huggingface_hub import login
|
| 12 |
from pydantic import BaseModel
|
|
|
|
| 19 |
from qwen_classifier.model import QwenClassifier
|
| 20 |
from qwen_classifier.config import HF_REPO, DEVICE
|
| 21 |
|
| 22 |
+
print(numpy.__version__)
|
| 23 |
|
| 24 |
app = FastAPI(title="Qwen Classifier")
|
| 25 |
hf_repo = os.getenv("HF_REPO")
|
qwen_classifier/evaluate.py
CHANGED
|
@@ -15,7 +15,7 @@ from pathlib import Path
|
|
| 15 |
from .config import TAG_NAMES, DEVICE, SPACE_URL, EVAL_LIMIT
|
| 16 |
from .globals import global_model, global_tokenizer
|
| 17 |
|
| 18 |
-
|
| 19 |
|
| 20 |
def _load_data(test_data_path):
|
| 21 |
eval_limit = os.getenv("EVAL_LIM")
|
|
|
|
| 15 |
from .config import TAG_NAMES, DEVICE, SPACE_URL, EVAL_LIMIT
|
| 16 |
from .globals import global_model, global_tokenizer
|
| 17 |
|
| 18 |
+
print(np.__version__)
|
| 19 |
|
| 20 |
def _load_data(test_data_path):
|
| 21 |
eval_limit = os.getenv("EVAL_LIM")
|
requirements.txt
CHANGED
|
@@ -1,5 +1,6 @@
|
|
| 1 |
# Core requirements from setup.py
|
| 2 |
torch>=2.1.0
|
|
|
|
| 3 |
transformers>=4.36.0
|
| 4 |
click>=8.1.0
|
| 5 |
scikit-learn>=1.3.0
|
|
|
|
| 1 |
# Core requirements from setup.py
|
| 2 |
torch>=2.1.0
|
| 3 |
+
numpy==1.23.5
|
| 4 |
transformers>=4.36.0
|
| 5 |
click>=8.1.0
|
| 6 |
scikit-learn>=1.3.0
|