Upload constants.py
Browse files- constants.py +49 -0
constants.py
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import os
|
2 |
+
|
3 |
+
# options
|
4 |
+
INDEX_BY_DESC = {
|
5 |
+
'OLMo 2 32B Instruct (4.6T tokens)': 'v4_olmo-2-0325-32b-instruct_llama',
|
6 |
+
'OLMo 2 13B Instruct (4.6T tokens)': 'v4_olmo-2-1124-13b-instruct_llama',
|
7 |
+
'OLMoE 1B 7B Instruct (4.6T tokens)': 'v4_olmoe-0125-1b-7b-instruct_llama',
|
8 |
+
'Dolma-v1.7 (2.6T tokens)': 'v4_dolma-v1_7_llama',
|
9 |
+
'RedPajama (1.4T tokens)': 'v4_rpj_llama_s4',
|
10 |
+
'Pile-train (380B tokens)': 'v4_piletrain_llama',
|
11 |
+
'C4-train (200B tokens)': 'v4_c4train_llama',
|
12 |
+
'Pile-val (390M tokens)': 'v4_pileval_llama',
|
13 |
+
# 'Pile-val (GPT-2 tokenizer), 380M tokens': 'v4_pileval_gpt2',
|
14 |
+
# 'Dolma-v1.6-sample (OLMo tokenizer), 8.0B tokens': 'v4_dolmasample_olmo',
|
15 |
+
# 'Dolma-v1.6-sample (9.2B tokens)': 'v4_dolma-v1_6-sample_llama',
|
16 |
+
# 'Dolma-v1.6 (3.1T tokens)': 'v4_dolma-v1_6_llama',
|
17 |
+
# 'Dolma-v1.6-wiki (4.3B tokens)': 'v4_dolma-v1_6-wiki_llama',
|
18 |
+
# 'Dolma-v1.6-books (5.8B tokens)': 'v4_dolma-v1_6-books_llama',
|
19 |
+
# 'Dolma-v1.6-pes2o (69B tokens)': 'v4_dolma-v1_6-pes2o_llama',
|
20 |
+
# 'Dolma-v1.6-reddit (89B tokens)': 'v4_dolma-v1_6-reddit_llama',
|
21 |
+
# 'Dolma-v1.6-c4 (200B tokens)': 'v4_dolma-v1_6-c4_llama',
|
22 |
+
# 'Dolma-v1.6-stack (420B tokens)': 'v4_dolma-v1_6-stack_llama',
|
23 |
+
# 'Dolma-v1.6-cc_en_head (660B tokens): 'v4_dolma-v1_6-cc_en_head_llama',
|
24 |
+
# 'Dolma-v1.6-cc_en_middle (650B tokens): 'v4_dolma-v1_6-cc_en_middle_llama',
|
25 |
+
# 'Dolma-v1.6-cc_en_tail (970B tokens): 'v4_dolma-v1_6-cc_en_tail_llama',
|
26 |
+
}
|
27 |
+
INDEX_DESCS = list(INDEX_BY_DESC.keys())
|
28 |
+
|
29 |
+
# API limits and defaults
|
30 |
+
MAX_QUERY_CHARS = int(os.environ.get('MAX_QUERY_CHARS', 1000))
|
31 |
+
MAX_CLAUSES_PER_CNF = int(os.environ.get('MAX_CLAUSES_PER_CNF', 4))
|
32 |
+
MAX_TERMS_PER_CLAUSE = int(os.environ.get('MAX_TERMS_PER_CLAUSE', 4))
|
33 |
+
max_support = int(os.environ.get('max_support', 1000))
|
34 |
+
MAX_SUPPORT = int(os.environ.get('MAX_SUPPORT', 1000))
|
35 |
+
max_clause_freq = int(os.environ.get('max_clause_freq', 50000))
|
36 |
+
MAX_CLAUSE_FREQ = int(os.environ.get('MAX_CLAUSE_FREQ', 500000))
|
37 |
+
max_diff_tokens = int(os.environ.get('max_diff_tokens', 100))
|
38 |
+
MAX_DIFF_TOKENS = int(os.environ.get('MAX_DIFF_TOKENS', 1000))
|
39 |
+
maxnum = int(os.environ.get('maxnum', 1))
|
40 |
+
MAXNUM = int(os.environ.get('MAXNUM', 10))
|
41 |
+
max_disp_len = int(os.environ.get('max_disp_len', 1000))
|
42 |
+
MAX_DISP_LEN = int(os.environ.get('MAX_DISP_LEN', 10000))
|
43 |
+
|
44 |
+
# HF demo
|
45 |
+
API_URL = os.environ.get('API_URL', None)
|
46 |
+
DEFAULT_CONCURRENCY_LIMIT = os.environ.get('DEFAULT_CONCURRENCY_LIMIT', 10)
|
47 |
+
MAX_SIZE = os.environ.get('MAX_SIZE', 100)
|
48 |
+
MAX_THREADS = os.environ.get('MAX_THREADS', 40)
|
49 |
+
DEBUG = (os.environ.get('DEBUG', 'False') != 'False')
|