import requests import re import json import time import os def extract_current_project(text): # 1. Locate currentProject start position key = '"currentProject":' start = text.find(key) if start == -1: print("currentProject not found!") return None start = start + len(key) # 2. Starting from the first {, use bracket counting method to find matching } while start < len(text) and text[start] not in '{': start += 1 if start == len(text): print("currentProject JSON start not found!") return None brace_count = 0 end = start for i, c in enumerate(text[start:]): if c == '{': brace_count += 1 elif c == '}': brace_count -= 1 if brace_count == 0: end = start + i + 1 break json_str = text[start:end] try: profile = json.loads(json_str) return profile except Exception as e: print(f"JSON decode error: {e}") return None def request_server_detail(url, headers): try: resp = requests.get(url, headers=headers, timeout=30) print(f"Status code: {resp.status_code} for {url}") if resp.status_code == 200: profile = extract_current_project(resp.text) return profile else: print(f"Failed to get detail: HTTP {resp.status_code}") return None except Exception as e: print(f"Exception: {e}") return None def batch_request_servers(): # Read mcpso_clients.json servers_path = os.path.join(os.path.dirname(__file__), 'mcpso_clients.json') with open(servers_path, 'r', encoding='utf-8') as f: servers = json.load(f) headers = { "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36", "Accept": "*/*", "Accept-Encoding": "gzip, deflate, br, zstd", "Accept-Language": "zh-CN,zh;q=0.9", "Cookie": "Your Cookie", "Next-Url": "/en/servers", "Priority": "u=1, i", "Referer": "https://mcp.so/servers", "Rsc": "1", "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', "Sec-Ch-Ua-Arch": "arm", "Sec-Ch-Ua-Bitness": "64", "Sec-Ch-Ua-Full-Version": "136.0.7103.114", "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"', "Sec-Ch-Ua-Mobile": "?0", "Sec-Ch-Ua-Model": '""', "Sec-Ch-Ua-Platform": '"macOS"', "Sec-Ch-Ua-Platform-Version": '"15.3.0"', "Sec-Fetch-Dest": "empty", "Sec-Fetch-Mode": "cors", "Sec-Fetch-Site": "same-origin" } for idx, item in enumerate(servers): # Skip already collected items (already have name field) if 'name' in item and 'metadata' in item: continue href = item.get('href') if not href: continue detail_url = f"https://mcp.so{href}" print(f"Requesting: {detail_url}") profile = request_server_detail(detail_url, headers) if not profile: print(f"Skip {href} due to extraction failure.") continue name = profile.get('name') url = profile.get('url') metadata = profile.copy() metadata.pop('name', None) metadata.pop('url', None) item['name'] = name item['url'] = url item['metadata'] = metadata # Write back in real time with open(servers_path, 'w', encoding='utf-8') as f: json.dump(servers, f, ensure_ascii=False, indent=2) print(f"Updated {idx+1}/{len(servers)}: {name}") time.sleep(1) print(f"All servers updated in {servers_path}") if __name__ == "__main__": # # Single collection (original logic) # url = "https://mcp.so/server/zhipu-web-search/BigModel?_rsc=n713a" # headers = { # "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36", # "Accept": "*/*", # "Accept-Encoding": "gzip, deflate, br, zstd", # "Accept-Language": "zh-CN,zh;q=0.9", # "Next-Url": "/en/server/zhipu-web-search/BigModel", # "Priority": "u=1, i", # "Referer": "https://mcp.so/server/zhipu-web-search/BigModel", # "Rsc": "1", # "Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', # "Sec-Ch-Ua-Arch": "arm", # "Sec-Ch-Ua-Bitness": "64", # "Sec-Ch-Ua-Full-Version": "136.0.7103.114", # "Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"', # "Sec-Ch-Ua-Mobile": "?0", # "Sec-Ch-Ua-Model": '""', # "Sec-Ch-Ua-Platform": '"macOS"', # "Sec-Ch-Ua-Platform-Version": '"15.3.0"', # "Sec-Fetch-Dest": "empty", # "Sec-Fetch-Mode": "cors", # "Sec-Fetch-Site": "same-origin" # } # profile = request_server_detail(url, headers) # if profile: # with open("server_zhipu-web-search_BigModel_profile.json", "w", encoding="utf-8") as f: # json.dump(profile, f, ensure_ascii=False, indent=2) # print("Profile saved to server_zhipu-web-search_BigModel_profile.json") # else: # print("Profile extraction failed!") # Batch collection batch_request_servers()