Update main.py
Browse files
main.py
CHANGED
@@ -16,14 +16,14 @@ def average_pool(last_hidden_states: Tensor,
|
|
16 |
|
17 |
# text-ada replacement
|
18 |
embeddingTokenizer = AutoTokenizer.from_pretrained(
|
19 |
-
'./multilingual-e5-base')
|
20 |
-
embeddingModel = AutoModel.from_pretrained('./multilingual-e5-base')
|
21 |
|
22 |
# chatGpt replacement
|
23 |
inferenceTokenizer = AutoTokenizer.from_pretrained(
|
24 |
-
"./ct2fast-flan-alpaca-xl")
|
25 |
inferenceTranslator = Translator(
|
26 |
-
"./ct2fast-flan-alpaca-xl", compute_type="int8", device="cpu")
|
27 |
|
28 |
|
29 |
class EmbeddingRequest(BaseModel):
|
@@ -99,7 +99,7 @@ async def tokens_count(request: TokensCountRequest):
|
|
99 |
inferenceTokenizer.encode(input_text))
|
100 |
|
101 |
# create response
|
102 |
-
|
103 |
'tokens': tokens,
|
104 |
'total': len(tokens)
|
105 |
}
|
|
|
16 |
|
17 |
# text-ada replacement
|
18 |
embeddingTokenizer = AutoTokenizer.from_pretrained(
|
19 |
+
'./models/multilingual-e5-base')
|
20 |
+
embeddingModel = AutoModel.from_pretrained('./models/multilingual-e5-base')
|
21 |
|
22 |
# chatGpt replacement
|
23 |
inferenceTokenizer = AutoTokenizer.from_pretrained(
|
24 |
+
"./models/ct2fast-flan-alpaca-xl")
|
25 |
inferenceTranslator = Translator(
|
26 |
+
"./models/ct2fast-flan-alpaca-xl", compute_type="int8", device="cpu")
|
27 |
|
28 |
|
29 |
class EmbeddingRequest(BaseModel):
|
|
|
99 |
inferenceTokenizer.encode(input_text))
|
100 |
|
101 |
# create response
|
102 |
+
return {
|
103 |
'tokens': tokens,
|
104 |
'total': len(tokens)
|
105 |
}
|