Datasets:
ArXiv:
License:
import requests | |
import re | |
import time | |
import os | |
import json | |
RETRY = 3 | |
SLEEP_BETWEEN_RETRY = 2 | |
def extract_hrefs(response_text): | |
# Extract all hrefs | |
return re.findall(r'"href":"(/client/[^"]+)"', response_text) | |
def main(): | |
output_file = "mcpso_all_hrefs.json" | |
visited = set() | |
href_list = [] | |
# Resume: load already saved hrefs | |
if os.path.exists(output_file): | |
with open(output_file, "r", encoding="utf-8") as f: | |
try: | |
href_list = json.load(f) | |
for item in href_list: | |
visited.add(item["href"]) | |
except Exception: | |
pass | |
headers = { | |
"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36", | |
"Accept": "*/*", | |
"Accept-Encoding": "gzip, deflate, br, zstd", | |
"Accept-Language": "zh-CN,zh;q=0.9", | |
"Cookie": "Your Cookie", | |
"Next-Url": "/en/servers", | |
"Priority": "u=1, i", | |
"Referer": "https://mcp.so/servers", | |
"Rsc": "1", | |
"Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', | |
"Sec-Ch-Ua-Arch": "arm", | |
"Sec-Ch-Ua-Bitness": "64", | |
"Sec-Ch-Ua-Full-Version": "136.0.7103.114", | |
"Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"', | |
"Sec-Ch-Ua-Mobile": "?0", | |
"Sec-Ch-Ua-Model": '""', | |
"Sec-Ch-Ua-Platform": '"macOS"', | |
"Sec-Ch-Ua-Platform-Version": '"15.3.0"', | |
"Sec-Fetch-Dest": "empty", | |
"Sec-Fetch-Mode": "cors", | |
"Sec-Fetch-Site": "same-origin" | |
} | |
for page in range(1, 477): | |
url = f"https://mcp.so/clients?page={page}" | |
print(f"Requesting page {page}: {url}") | |
for attempt in range(1, RETRY + 1): | |
try: | |
resp = requests.get(url, headers=headers, timeout=30) | |
if resp.status_code != 200: | |
print(f"Page {page} failed: HTTP {resp.status_code}, attempt {attempt}/{RETRY}") | |
time.sleep(SLEEP_BETWEEN_RETRY) | |
continue | |
hrefs = extract_hrefs(resp.text) | |
new_hrefs = [h for h in hrefs if h not in visited] | |
for h in new_hrefs: | |
href_list.append({"href": h}) | |
visited.add(h) | |
# Save in real time | |
with open(output_file, "w", encoding="utf-8") as f: | |
json.dump(href_list, f, ensure_ascii=False, indent=2) | |
print(f"Page {page} got {len(new_hrefs)} new, total {len(href_list)}") | |
time.sleep(1) | |
break | |
except Exception as e: | |
print(f"Page {page} exception: {e}, attempt {attempt}/{RETRY}") | |
time.sleep(SLEEP_BETWEEN_RETRY) | |
continue | |
else: | |
print(f"Page {page} failed after {RETRY} retries.") | |
print(f"All done. Total unique hrefs: {len(href_list)}. Saved to {output_file}") | |
if __name__ == "__main__": | |
main() |