Datasets:
ArXiv:
License:
File size: 5,541 Bytes
e243320 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 |
import requests
import re
import json
import time
import os
def extract_current_project(text):
# 1. Locate currentProject start position
key = '"currentProject":'
start = text.find(key)
if start == -1:
print("currentProject not found!")
return None
start = start + len(key)
# 2. Starting from the first {, use bracket counting method to find matching }
while start < len(text) and text[start] not in '{':
start += 1
if start == len(text):
print("currentProject JSON start not found!")
return None
brace_count = 0
end = start
for i, c in enumerate(text[start:]):
if c == '{':
brace_count += 1
elif c == '}':
brace_count -= 1
if brace_count == 0:
end = start + i + 1
break
json_str = text[start:end]
try:
profile = json.loads(json_str)
return profile
except Exception as e:
print(f"JSON decode error: {e}")
return None
def request_server_detail(url, headers):
try:
resp = requests.get(url, headers=headers, timeout=30)
print(f"Status code: {resp.status_code} for {url}")
if resp.status_code == 200:
profile = extract_current_project(resp.text)
return profile
else:
print(f"Failed to get detail: HTTP {resp.status_code}")
return None
except Exception as e:
print(f"Exception: {e}")
return None
def batch_request_servers():
# Read mcpso_servers.json
servers_path = os.path.join(os.path.dirname(__file__), 'mcpso_servers.json')
with open(servers_path, 'r', encoding='utf-8') as f:
servers = json.load(f)
headers = {
"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
"Accept": "*/*",
"Accept-Encoding": "gzip, deflate, br, zstd",
"Accept-Language": "zh-CN,zh;q=0.9",
"Cookie": "Your Cookie",
"Next-Url": "/en/server/zhipu-web-search/BigModel",
"Priority": "u=1, i",
"Referer": "https://mcp.so/server/zhipu-web-search/BigModel",
"Rsc": "1",
"Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
"Sec-Ch-Ua-Arch": "arm",
"Sec-Ch-Ua-Bitness": "64",
"Sec-Ch-Ua-Full-Version": "136.0.7103.114",
"Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
"Sec-Ch-Ua-Mobile": "?0",
"Sec-Ch-Ua-Model": '""',
"Sec-Ch-Ua-Platform": '"macOS"',
"Sec-Ch-Ua-Platform-Version": '"15.3.0"',
"Sec-Fetch-Dest": "empty",
"Sec-Fetch-Mode": "cors",
"Sec-Fetch-Site": "same-origin"
}
for idx, item in enumerate(servers):
# Skip already collected items (already have name field)
if 'name' in item and 'metadata' in item:
continue
href = item.get('href')
if not href:
continue
detail_url = f"https://mcp.so{href}"
print(f"Requesting: {detail_url}")
profile = request_server_detail(detail_url, headers)
if not profile:
print(f"Skip {href} due to extraction failure.")
continue
name = profile.get('name')
url = profile.get('url')
metadata = profile.copy()
metadata.pop('name', None)
metadata.pop('url', None)
item['name'] = name
item['url'] = url
item['metadata'] = metadata
# Write back in real time
with open(servers_path, 'w', encoding='utf-8') as f:
json.dump(servers, f, ensure_ascii=False, indent=2)
print(f"Updated {idx+1}/{len(servers)}: {name}")
time.sleep(1)
print(f"All servers updated in {servers_path}")
if __name__ == "__main__":
#
# url = "https://mcp.so/server/zhipu-web-search/BigModel?_rsc=n713a"
# headers = {
# "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
# "Accept": "*/*",
# "Accept-Encoding": "gzip, deflate, br, zstd",
# "Accept-Language": "zh-CN,zh;q=0.9",
# "Next-Url": "/en/server/zhipu-web-search/BigModel",
# "Priority": "u=1, i",
# "Referer": "https://mcp.so/server/zhipu-web-search/BigModel",
# "Rsc": "1",
# "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
# "Sec-Ch-Ua-Arch": "arm",
# "Sec-Ch-Ua-Bitness": "64",
# "Sec-Ch-Ua-Full-Version": "136.0.7103.114",
# "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
# "Sec-Ch-Ua-Mobile": "?0",
# "Sec-Ch-Ua-Model": '""',
# "Sec-Ch-Ua-Platform": '"macOS"',
# "Sec-Ch-Ua-Platform-Version": '"15.3.0"',
# "Sec-Fetch-Dest": "empty",
# "Sec-Fetch-Mode": "cors",
# "Sec-Fetch-Site": "same-origin"
# }
# profile = request_server_detail(url, headers)
# if profile:
# with open("server_zhipu-web-search_BigModel_profile.json", "w", encoding="utf-8") as f:
# json.dump(profile, f, ensure_ascii=False, indent=2)
# print("Profile saved to server_zhipu-web-search_BigModel_profile.json")
# else:
# print("Profile extraction failed!")
#
batch_request_servers() |