Snak1nya commited on
Commit
e243320
·
verified ·
1 Parent(s): 3036f5b

Upload folder using huggingface_hub

Browse files
Crawler/Clients/Client_request.py ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+ import re
3
+ import time
4
+ import os
5
+ import json
6
+
7
+ RETRY = 3
8
+ SLEEP_BETWEEN_RETRY = 2
9
+
10
+ def extract_hrefs(response_text):
11
+ # Extract all hrefs
12
+ return re.findall(r'"href":"(/client/[^"]+)"', response_text)
13
+
14
+ def main():
15
+ output_file = "mcpso_all_hrefs.json"
16
+ visited = set()
17
+ href_list = []
18
+ # Resume: load already saved hrefs
19
+ if os.path.exists(output_file):
20
+ with open(output_file, "r", encoding="utf-8") as f:
21
+ try:
22
+ href_list = json.load(f)
23
+ for item in href_list:
24
+ visited.add(item["href"])
25
+ except Exception:
26
+ pass
27
+
28
+ headers = {
29
+ "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
30
+ "Accept": "*/*",
31
+ "Accept-Encoding": "gzip, deflate, br, zstd",
32
+ "Accept-Language": "zh-CN,zh;q=0.9",
33
+ "Cookie": "Your Cookie",
34
+ "Next-Url": "/en/servers",
35
+ "Priority": "u=1, i",
36
+ "Referer": "https://mcp.so/servers",
37
+ "Rsc": "1",
38
+ "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
39
+ "Sec-Ch-Ua-Arch": "arm",
40
+ "Sec-Ch-Ua-Bitness": "64",
41
+ "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
42
+ "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
43
+ "Sec-Ch-Ua-Mobile": "?0",
44
+ "Sec-Ch-Ua-Model": '""',
45
+ "Sec-Ch-Ua-Platform": '"macOS"',
46
+ "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
47
+ "Sec-Fetch-Dest": "empty",
48
+ "Sec-Fetch-Mode": "cors",
49
+ "Sec-Fetch-Site": "same-origin"
50
+ }
51
+
52
+ for page in range(1, 477):
53
+ url = f"https://mcp.so/clients?page={page}"
54
+ print(f"Requesting page {page}: {url}")
55
+ for attempt in range(1, RETRY + 1):
56
+ try:
57
+ resp = requests.get(url, headers=headers, timeout=30)
58
+ if resp.status_code != 200:
59
+ print(f"Page {page} failed: HTTP {resp.status_code}, attempt {attempt}/{RETRY}")
60
+ time.sleep(SLEEP_BETWEEN_RETRY)
61
+ continue
62
+ hrefs = extract_hrefs(resp.text)
63
+ new_hrefs = [h for h in hrefs if h not in visited]
64
+ for h in new_hrefs:
65
+ href_list.append({"href": h})
66
+ visited.add(h)
67
+ # Save in real time
68
+ with open(output_file, "w", encoding="utf-8") as f:
69
+ json.dump(href_list, f, ensure_ascii=False, indent=2)
70
+ print(f"Page {page} got {len(new_hrefs)} new, total {len(href_list)}")
71
+ time.sleep(1)
72
+ break
73
+ except Exception as e:
74
+ print(f"Page {page} exception: {e}, attempt {attempt}/{RETRY}")
75
+ time.sleep(SLEEP_BETWEEN_RETRY)
76
+ continue
77
+ else:
78
+ print(f"Page {page} failed after {RETRY} retries.")
79
+ print(f"All done. Total unique hrefs: {len(href_list)}. Saved to {output_file}")
80
+
81
+ if __name__ == "__main__":
82
+ main()
Crawler/Clients/mcpso_client_detail_request.py ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+ import re
3
+ import json
4
+ import time
5
+ import os
6
+
7
+ def extract_current_project(text):
8
+ # 1. Locate currentProject start position
9
+ key = '"currentProject":'
10
+ start = text.find(key)
11
+ if start == -1:
12
+ print("currentProject not found!")
13
+ return None
14
+ start = start + len(key)
15
+ # 2. Starting from the first {, use bracket counting method to find matching }
16
+ while start < len(text) and text[start] not in '{':
17
+ start += 1
18
+ if start == len(text):
19
+ print("currentProject JSON start not found!")
20
+ return None
21
+ brace_count = 0
22
+ end = start
23
+ for i, c in enumerate(text[start:]):
24
+ if c == '{':
25
+ brace_count += 1
26
+ elif c == '}':
27
+ brace_count -= 1
28
+ if brace_count == 0:
29
+ end = start + i + 1
30
+ break
31
+ json_str = text[start:end]
32
+ try:
33
+ profile = json.loads(json_str)
34
+ return profile
35
+ except Exception as e:
36
+ print(f"JSON decode error: {e}")
37
+ return None
38
+
39
+ def request_server_detail(url, headers):
40
+ try:
41
+ resp = requests.get(url, headers=headers, timeout=30)
42
+ print(f"Status code: {resp.status_code} for {url}")
43
+ if resp.status_code == 200:
44
+ profile = extract_current_project(resp.text)
45
+ return profile
46
+ else:
47
+ print(f"Failed to get detail: HTTP {resp.status_code}")
48
+ return None
49
+ except Exception as e:
50
+ print(f"Exception: {e}")
51
+ return None
52
+
53
+ def batch_request_servers():
54
+ # Read mcpso_clients.json
55
+ servers_path = os.path.join(os.path.dirname(__file__), 'mcpso_clients.json')
56
+ with open(servers_path, 'r', encoding='utf-8') as f:
57
+ servers = json.load(f)
58
+ headers = {
59
+ "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
60
+ "Accept": "*/*",
61
+ "Accept-Encoding": "gzip, deflate, br, zstd",
62
+ "Accept-Language": "zh-CN,zh;q=0.9",
63
+ "Cookie": "Your Cookie",
64
+ "Next-Url": "/en/servers",
65
+ "Priority": "u=1, i",
66
+ "Referer": "https://mcp.so/servers",
67
+ "Rsc": "1",
68
+ "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
69
+ "Sec-Ch-Ua-Arch": "arm",
70
+ "Sec-Ch-Ua-Bitness": "64",
71
+ "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
72
+ "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
73
+ "Sec-Ch-Ua-Mobile": "?0",
74
+ "Sec-Ch-Ua-Model": '""',
75
+ "Sec-Ch-Ua-Platform": '"macOS"',
76
+ "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
77
+ "Sec-Fetch-Dest": "empty",
78
+ "Sec-Fetch-Mode": "cors",
79
+ "Sec-Fetch-Site": "same-origin"
80
+ }
81
+ for idx, item in enumerate(servers):
82
+ # Skip already collected items (already have name field)
83
+ if 'name' in item and 'metadata' in item:
84
+ continue
85
+ href = item.get('href')
86
+ if not href:
87
+ continue
88
+ detail_url = f"https://mcp.so{href}"
89
+ print(f"Requesting: {detail_url}")
90
+ profile = request_server_detail(detail_url, headers)
91
+ if not profile:
92
+ print(f"Skip {href} due to extraction failure.")
93
+ continue
94
+ name = profile.get('name')
95
+ url = profile.get('url')
96
+ metadata = profile.copy()
97
+ metadata.pop('name', None)
98
+ metadata.pop('url', None)
99
+ item['name'] = name
100
+ item['url'] = url
101
+ item['metadata'] = metadata
102
+ # Write back in real time
103
+ with open(servers_path, 'w', encoding='utf-8') as f:
104
+ json.dump(servers, f, ensure_ascii=False, indent=2)
105
+ print(f"Updated {idx+1}/{len(servers)}: {name}")
106
+ time.sleep(1)
107
+ print(f"All servers updated in {servers_path}")
108
+
109
+ if __name__ == "__main__":
110
+ # # Single collection (original logic)
111
+ # url = "https://mcp.so/server/zhipu-web-search/BigModel?_rsc=n713a"
112
+ # headers = {
113
+ # "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
114
+ # "Accept": "*/*",
115
+ # "Accept-Encoding": "gzip, deflate, br, zstd",
116
+ # "Accept-Language": "zh-CN,zh;q=0.9",
117
+ # "Next-Url": "/en/server/zhipu-web-search/BigModel",
118
+ # "Priority": "u=1, i",
119
+ # "Referer": "https://mcp.so/server/zhipu-web-search/BigModel",
120
+ # "Rsc": "1",
121
+ # "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
122
+ # "Sec-Ch-Ua-Arch": "arm",
123
+ # "Sec-Ch-Ua-Bitness": "64",
124
+ # "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
125
+ # "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
126
+ # "Sec-Ch-Ua-Mobile": "?0",
127
+ # "Sec-Ch-Ua-Model": '""',
128
+ # "Sec-Ch-Ua-Platform": '"macOS"',
129
+ # "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
130
+ # "Sec-Fetch-Dest": "empty",
131
+ # "Sec-Fetch-Mode": "cors",
132
+ # "Sec-Fetch-Site": "same-origin"
133
+ # }
134
+ # profile = request_server_detail(url, headers)
135
+ # if profile:
136
+ # with open("server_zhipu-web-search_BigModel_profile.json", "w", encoding="utf-8") as f:
137
+ # json.dump(profile, f, ensure_ascii=False, indent=2)
138
+ # print("Profile saved to server_zhipu-web-search_BigModel_profile.json")
139
+ # else:
140
+ # print("Profile extraction failed!")
141
+ # Batch collection
142
+ batch_request_servers()
Crawler/Clients/mcpso_clients.json ADDED
The diff for this file is too large to render. See raw diff
 
Crawler/Servers/Server_request.py ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+ import re
3
+ import time
4
+ import os
5
+ import json
6
+
7
+ RETRY = 3
8
+ SLEEP_BETWEEN_RETRY = 2
9
+
10
+ def extract_hrefs(response_text):
11
+ # Extract all hrefs
12
+ return re.findall(r'"href":"(/server/[^"]+)"', response_text)
13
+
14
+ def main():
15
+ output_file = "mcpso_all_hrefs.json"
16
+ visited = set()
17
+ href_list = []
18
+ # Resume: load already saved hrefs
19
+ if os.path.exists(output_file):
20
+ with open(output_file, "r", encoding="utf-8") as f:
21
+ try:
22
+ href_list = json.load(f)
23
+ for item in href_list:
24
+ visited.add(item["href"])
25
+ except Exception:
26
+ pass
27
+
28
+ headers = {
29
+ "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
30
+ "Accept": "*/*",
31
+ "Accept-Encoding": "gzip, deflate, br, zstd",
32
+ "Accept-Language": "zh-CN,zh;q=0.9",
33
+ "Cookie": "Your Cookie",
34
+ "Next-Url": "/en/servers",
35
+ "Priority": "u=1, i",
36
+ "Referer": "https://mcp.so/servers",
37
+ "Rsc": "1",
38
+ "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
39
+ "Sec-Ch-Ua-Arch": "arm",
40
+ "Sec-Ch-Ua-Bitness": "64",
41
+ "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
42
+ "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
43
+ "Sec-Ch-Ua-Mobile": "?0",
44
+ "Sec-Ch-Ua-Model": '""',
45
+ "Sec-Ch-Ua-Platform": '"macOS"',
46
+ "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
47
+ "Sec-Fetch-Dest": "empty",
48
+ "Sec-Fetch-Mode": "cors",
49
+ "Sec-Fetch-Site": "same-origin"
50
+ }
51
+
52
+ for page in range(1, 477):
53
+ url = f"https://mcp.so/servers?page={page}"
54
+ print(f"Requesting page {page}: {url}")
55
+ for attempt in range(1, RETRY + 1):
56
+ try:
57
+ resp = requests.get(url, headers=headers, timeout=30)
58
+ if resp.status_code != 200:
59
+ print(f"Page {page} failed: HTTP {resp.status_code}, attempt {attempt}/{RETRY}")
60
+ time.sleep(SLEEP_BETWEEN_RETRY)
61
+ continue
62
+ hrefs = extract_hrefs(resp.text)
63
+ new_hrefs = [h for h in hrefs if h not in visited]
64
+ for h in new_hrefs:
65
+ href_list.append({"href": h})
66
+ visited.add(h)
67
+ # Save in real time
68
+ with open(output_file, "w", encoding="utf-8") as f:
69
+ json.dump(href_list, f, ensure_ascii=False, indent=2)
70
+ print(f"Page {page} got {len(new_hrefs)} new, total {len(href_list)}")
71
+ time.sleep(1)
72
+ break
73
+ except Exception as e:
74
+ print(f"Page {page} exception: {e}, attempt {attempt}/{RETRY}")
75
+ time.sleep(SLEEP_BETWEEN_RETRY)
76
+ continue
77
+ else:
78
+ print(f"Page {page} failed after {RETRY} retries.")
79
+ print(f"All done. Total unique hrefs: {len(href_list)}. Saved to {output_file}")
80
+
81
+ if __name__ == "__main__":
82
+ main()
Crawler/Servers/mcpso_server_detail_request.py ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+ import re
3
+ import json
4
+ import time
5
+ import os
6
+
7
+ def extract_current_project(text):
8
+ # 1. Locate currentProject start position
9
+ key = '"currentProject":'
10
+ start = text.find(key)
11
+ if start == -1:
12
+ print("currentProject not found!")
13
+ return None
14
+ start = start + len(key)
15
+ # 2. Starting from the first {, use bracket counting method to find matching }
16
+ while start < len(text) and text[start] not in '{':
17
+ start += 1
18
+ if start == len(text):
19
+ print("currentProject JSON start not found!")
20
+ return None
21
+ brace_count = 0
22
+ end = start
23
+ for i, c in enumerate(text[start:]):
24
+ if c == '{':
25
+ brace_count += 1
26
+ elif c == '}':
27
+ brace_count -= 1
28
+ if brace_count == 0:
29
+ end = start + i + 1
30
+ break
31
+ json_str = text[start:end]
32
+ try:
33
+ profile = json.loads(json_str)
34
+ return profile
35
+ except Exception as e:
36
+ print(f"JSON decode error: {e}")
37
+ return None
38
+
39
+ def request_server_detail(url, headers):
40
+ try:
41
+ resp = requests.get(url, headers=headers, timeout=30)
42
+ print(f"Status code: {resp.status_code} for {url}")
43
+ if resp.status_code == 200:
44
+ profile = extract_current_project(resp.text)
45
+ return profile
46
+ else:
47
+ print(f"Failed to get detail: HTTP {resp.status_code}")
48
+ return None
49
+ except Exception as e:
50
+ print(f"Exception: {e}")
51
+ return None
52
+
53
+ def batch_request_servers():
54
+ # Read mcpso_servers.json
55
+ servers_path = os.path.join(os.path.dirname(__file__), 'mcpso_servers.json')
56
+ with open(servers_path, 'r', encoding='utf-8') as f:
57
+ servers = json.load(f)
58
+ headers = {
59
+ "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
60
+ "Accept": "*/*",
61
+ "Accept-Encoding": "gzip, deflate, br, zstd",
62
+ "Accept-Language": "zh-CN,zh;q=0.9",
63
+ "Cookie": "Your Cookie",
64
+ "Next-Url": "/en/server/zhipu-web-search/BigModel",
65
+ "Priority": "u=1, i",
66
+ "Referer": "https://mcp.so/server/zhipu-web-search/BigModel",
67
+ "Rsc": "1",
68
+ "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
69
+ "Sec-Ch-Ua-Arch": "arm",
70
+ "Sec-Ch-Ua-Bitness": "64",
71
+ "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
72
+ "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
73
+ "Sec-Ch-Ua-Mobile": "?0",
74
+ "Sec-Ch-Ua-Model": '""',
75
+ "Sec-Ch-Ua-Platform": '"macOS"',
76
+ "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
77
+ "Sec-Fetch-Dest": "empty",
78
+ "Sec-Fetch-Mode": "cors",
79
+ "Sec-Fetch-Site": "same-origin"
80
+ }
81
+ for idx, item in enumerate(servers):
82
+ # Skip already collected items (already have name field)
83
+ if 'name' in item and 'metadata' in item:
84
+ continue
85
+ href = item.get('href')
86
+ if not href:
87
+ continue
88
+ detail_url = f"https://mcp.so{href}"
89
+ print(f"Requesting: {detail_url}")
90
+ profile = request_server_detail(detail_url, headers)
91
+ if not profile:
92
+ print(f"Skip {href} due to extraction failure.")
93
+ continue
94
+ name = profile.get('name')
95
+ url = profile.get('url')
96
+ metadata = profile.copy()
97
+ metadata.pop('name', None)
98
+ metadata.pop('url', None)
99
+ item['name'] = name
100
+ item['url'] = url
101
+ item['metadata'] = metadata
102
+ # Write back in real time
103
+ with open(servers_path, 'w', encoding='utf-8') as f:
104
+ json.dump(servers, f, ensure_ascii=False, indent=2)
105
+ print(f"Updated {idx+1}/{len(servers)}: {name}")
106
+ time.sleep(1)
107
+ print(f"All servers updated in {servers_path}")
108
+
109
+ if __name__ == "__main__":
110
+ #
111
+ # url = "https://mcp.so/server/zhipu-web-search/BigModel?_rsc=n713a"
112
+ # headers = {
113
+ # "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
114
+ # "Accept": "*/*",
115
+ # "Accept-Encoding": "gzip, deflate, br, zstd",
116
+ # "Accept-Language": "zh-CN,zh;q=0.9",
117
+ # "Next-Url": "/en/server/zhipu-web-search/BigModel",
118
+ # "Priority": "u=1, i",
119
+ # "Referer": "https://mcp.so/server/zhipu-web-search/BigModel",
120
+ # "Rsc": "1",
121
+ # "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
122
+ # "Sec-Ch-Ua-Arch": "arm",
123
+ # "Sec-Ch-Ua-Bitness": "64",
124
+ # "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
125
+ # "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
126
+ # "Sec-Ch-Ua-Mobile": "?0",
127
+ # "Sec-Ch-Ua-Model": '""',
128
+ # "Sec-Ch-Ua-Platform": '"macOS"',
129
+ # "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
130
+ # "Sec-Fetch-Dest": "empty",
131
+ # "Sec-Fetch-Mode": "cors",
132
+ # "Sec-Fetch-Site": "same-origin"
133
+ # }
134
+ # profile = request_server_detail(url, headers)
135
+ # if profile:
136
+ # with open("server_zhipu-web-search_BigModel_profile.json", "w", encoding="utf-8") as f:
137
+ # json.dump(profile, f, ensure_ascii=False, indent=2)
138
+ # print("Profile saved to server_zhipu-web-search_BigModel_profile.json")
139
+ # else:
140
+ # print("Profile extraction failed!")
141
+ #
142
+ batch_request_servers()
Crawler/Servers/mcpso_servers.json ADDED
The diff for this file is too large to render. See raw diff
 
Crawler/data_cleaner.py ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ # -*- coding: utf-8 -*-
3
+ """
4
+ MCP Servers Data Cleaning Script
5
+ Keep only user-specified core fields and simplify data structure
6
+ """
7
+
8
+ import json
9
+ import os
10
+ from typing import Dict, Any, List
11
+
12
+ def clean_server_data(original_data: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
13
+ """
14
+ Clean server data, keep only specified fields, and filter out items without URLs
15
+ """
16
+ cleaned_data = []
17
+ filtered_count = 0
18
+
19
+ for item in original_data:
20
+ # Filter out items without URLs
21
+ url = item.get('url')
22
+ if not url or url.strip() == '':
23
+ filtered_count += 1
24
+ continue
25
+
26
+ metadata = item.get('metadata', {})
27
+ github = item.get('github', {})
28
+
29
+ # Build cleaned data structure
30
+ cleaned_item = {
31
+ # Unique identifier
32
+ "id": metadata.get('id'),
33
+
34
+ # Basic information
35
+ "name": item.get('name'),
36
+ "url": url,
37
+
38
+ # metadata core fields
39
+ "title": metadata.get('title'),
40
+ "description": metadata.get('description'),
41
+ "author_name": metadata.get('author_name'),
42
+ "tags": metadata.get('tags'),
43
+ "category": metadata.get('category'),
44
+ "type": metadata.get('type'),
45
+ "tools": metadata.get('tools'),
46
+ "sse_url": metadata.get('sse_url'),
47
+ "server_command": metadata.get('server_command'),
48
+ "server_config": metadata.get('server_config'),
49
+
50
+ # github complete object
51
+ "github": github if github else None
52
+ }
53
+
54
+ cleaned_data.append(cleaned_item)
55
+
56
+ print(f"Filtered out items without URLs: {filtered_count} records")
57
+ return cleaned_data
58
+
59
+ def main():
60
+ """Main function"""
61
+ input_file = 'mcpso_servers.json'
62
+ output_file = 'mcpso_servers_cleaned.json'
63
+
64
+ # Check input file
65
+ if not os.path.exists(input_file):
66
+ print(f"❌ Input file not found: {input_file}")
67
+ return
68
+
69
+ # Read original data
70
+ print(f"📖 Reading original data: {input_file}")
71
+ with open(input_file, 'r', encoding='utf-8') as f:
72
+ original_data = json.load(f)
73
+
74
+ print(f"Original data count: {len(original_data)}")
75
+
76
+ # Clean data
77
+ print("🧹 Starting data cleaning...")
78
+ cleaned_data = clean_server_data(original_data)
79
+
80
+ # Statistics
81
+ github_count = sum(1 for item in cleaned_data if item.get('github'))
82
+ print(f"Cleaned data count: {len(cleaned_data)}")
83
+ print(f"Contains GitHub information: {github_count} records")
84
+
85
+ # Write cleaned data
86
+ print(f"💾 Writing cleaned data: {output_file}")
87
+ with open(output_file, 'w', encoding='utf-8') as f:
88
+ json.dump(cleaned_data, f, ensure_ascii=False, indent=2)
89
+
90
+ # Calculate file size change
91
+ original_size = os.path.getsize(input_file) / 1024 / 1024 # MB
92
+ cleaned_size = os.path.getsize(output_file) / 1024 / 1024 # MB
93
+ size_reduction = (1 - cleaned_size / original_size) * 100
94
+
95
+ print(f"\n📊 Cleaning Results:")
96
+ print(f"Original file size: {original_size:.2f} MB")
97
+ print(f"Cleaned file size: {cleaned_size:.2f} MB")
98
+ print(f"Size reduction: {size_reduction:.1f}%")
99
+ print(f"✅ Data cleaning completed!")
100
+
101
+ if __name__ == "__main__":
102
+ main()
Crawler/github_info_collector.py ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import re
3
+ import sys
4
+ import json
5
+ import time
6
+ import requests
7
+ from urllib.parse import urlparse
8
+
9
+ def robust_request(url, headers=None, max_retries=5, timeout=30):
10
+ """
11
+ Robust requests.get with retry logic and GitHub API rate limit handling.
12
+ """
13
+ delay = 5
14
+ for attempt in range(max_retries):
15
+ try:
16
+ resp = requests.get(url, headers=headers, timeout=timeout)
17
+ if resp.status_code == 403:
18
+ # Check if it's rate limit
19
+ if 'X-RateLimit-Remaining' in resp.headers and resp.headers['X-RateLimit-Remaining'] == '0':
20
+ reset_ts = int(resp.headers.get('X-RateLimit-Reset', time.time() + 60))
21
+ wait_sec = max(reset_ts - int(time.time()), 5)
22
+ print(f"[Rate Limit] API rate limit, waiting {wait_sec} seconds...")
23
+ time.sleep(wait_sec)
24
+ continue
25
+ else:
26
+ print(f"[WARN] 403 Forbidden: {url}")
27
+ time.sleep(delay)
28
+ delay = min(delay * 2, 120)
29
+ continue
30
+ elif resp.status_code in (429, 502, 503, 504):
31
+ print(f"[WARN] {resp.status_code}, retrying {url}, waiting {delay} seconds...")
32
+ time.sleep(delay)
33
+ delay = min(delay * 2, 120)
34
+ continue
35
+ return resp
36
+ except Exception as e:
37
+ print(f"[ERROR] Network exception: {e}, retrying {url}, waiting {delay} seconds...")
38
+ time.sleep(delay)
39
+ delay = min(delay * 2, 120)
40
+ print(f"[FATAL] Multiple retries failed: {url}")
41
+ return None
42
+
43
+ def extract_github_repo(url):
44
+ """Extract owner/repo from GitHub URL"""
45
+ if not url or 'github.com' not in url:
46
+ return None
47
+ m = re.search(r'github.com/([\w\-\.]+)/([\w\-\.]+)', url)
48
+ if m:
49
+ return f"{m.group(1)}/{m.group(2)}"
50
+ return None
51
+
52
+ def get_github_info(full_name, token=None):
53
+ """Get GitHub repository information"""
54
+ headers = {'Accept': 'application/vnd.github+json'}
55
+ if token:
56
+ headers['Authorization'] = f'token {token}'
57
+ api_url = f'https://api.github.com/repos/{full_name}'
58
+ repo_resp = robust_request(api_url, headers)
59
+ if not repo_resp or repo_resp.status_code != 200:
60
+ print(f"[WARN] Failed to get repository info: {full_name}")
61
+ return None
62
+ repo = repo_resp.json()
63
+ # Contributors count
64
+ contrib_url = f'https://api.github.com/repos/{full_name}/contributors?per_page=1&anon=true'
65
+ contrib_resp = robust_request(contrib_url, headers)
66
+ contributors_count = 0
67
+ if contrib_resp and contrib_resp.status_code == 200:
68
+ if 'Link' in contrib_resp.headers and 'last' in contrib_resp.headers['Link']:
69
+ last_link = contrib_resp.headers['Link'].split(',')[-1]
70
+ m = re.search(r'&page=(\d+)>; rel="last"', last_link)
71
+ if m:
72
+ contributors_count = int(m.group(1))
73
+ else:
74
+ contributors_count = len(contrib_resp.json())
75
+ # Language statistics
76
+ lang_url = f'https://api.github.com/repos/{full_name}/languages'
77
+ lang_resp = robust_request(lang_url, headers)
78
+ languages = lang_resp.json() if lang_resp and lang_resp.status_code == 200 else {}
79
+ # File detection
80
+ tree_url = f'https://api.github.com/repos/{full_name}/git/trees/{repo.get('default_branch', 'main')}?recursive=1'
81
+ tree_resp = robust_request(tree_url, headers)
82
+ has_docker = has_readme = has_requirements = False
83
+ if tree_resp and tree_resp.status_code == 200:
84
+ files = [item['path'].lower() for item in tree_resp.json().get('tree', []) if item['type'] == 'blob']
85
+ has_docker = any('dockerfile' in f for f in files)
86
+ has_readme = any(f.startswith('readme') for f in files)
87
+ has_requirements = any('requirements.txt' in f for f in files)
88
+ # Last commit
89
+ commit_url = f'https://api.github.com/repos/{full_name}/commits?per_page=1'
90
+ commit_resp = robust_request(commit_url, headers)
91
+ last_commit = None
92
+ if commit_resp and commit_resp.status_code == 200 and len(commit_resp.json()) > 0:
93
+ last_commit = commit_resp.json()[0]['commit']['committer']['date']
94
+ # license
95
+ license_name = repo['license']['name'] if repo.get('license') else None
96
+ return {
97
+ "full_name": full_name,
98
+ "stargazers_count": repo.get('stargazers_count', 0),
99
+ "forks_count": repo.get('forks_count', 0),
100
+ "open_issues_count": repo.get('open_issues_count', 0),
101
+ "contributors_count": contributors_count,
102
+ "language": repo.get('language'),
103
+ "languages": languages,
104
+ "license": license_name,
105
+ "archived": repo.get('archived', False),
106
+ "has_docker": has_docker,
107
+ "has_readme": has_readme,
108
+ "has_requirements": has_requirements,
109
+ "last_commit": last_commit
110
+ }
111
+
112
+ def update_json_file(json_path, token=None):
113
+ with open(json_path, 'r', encoding='utf-8') as f:
114
+ servers = json.load(f)
115
+ changed = False
116
+ for idx, item in enumerate(servers):
117
+ url = item.get('url')
118
+ if not url or 'github.com' not in url:
119
+ continue
120
+ if 'github' in item and item['github']:
121
+ continue # Already collected
122
+ full_name = extract_github_repo(url)
123
+ if not full_name:
124
+ continue
125
+ print(f"[{idx+1}/{len(servers)}] Collecting {full_name} ...")
126
+ info = get_github_info(full_name, token)
127
+ if info:
128
+ item['github'] = info
129
+ changed = True
130
+ # Write back in real time
131
+ with open(json_path, 'w', encoding='utf-8') as f:
132
+ json.dump(servers, f, ensure_ascii=False, indent=2)
133
+ time.sleep(0.5) # Prevent API rate limiting
134
+ if changed:
135
+ print(f"All collection completed, written back to {json_path}")
136
+ else:
137
+ print("No repositories need to be updated.")
138
+
139
+ def test_single_url(url, token=None):
140
+ full_name = extract_github_repo(url)
141
+ if not full_name:
142
+ print("Not a valid GitHub URL")
143
+ return
144
+ info = get_github_info(full_name, token)
145
+ print(json.dumps(info, ensure_ascii=False, indent=2))
146
+
147
+ if __name__ == "__main__":
148
+ import argparse
149
+ parser = argparse.ArgumentParser(description='Batch/single collection of GitHub repository information')
150
+ parser.add_argument('--file', type=str, default='mcpso_servers.json', help='JSON file path')
151
+ parser.add_argument('--url', type=str, help='Single GitHub repository URL')
152
+ parser.add_argument('--token', type=str, help='GitHub API Token (optional)')
153
+ args = parser.parse_args()
154
+ if args.url:
155
+ test_single_url(args.url, args.token)
156
+ else:
157
+ update_json_file(args.file, args.token)