MCPCorpus / Crawler /Clients /Client_request.py
Snak1nya's picture
Upload folder using huggingface_hub
e243320 verified
import requests
import re
import time
import os
import json
RETRY = 3
SLEEP_BETWEEN_RETRY = 2
def extract_hrefs(response_text):
# Extract all hrefs
return re.findall(r'"href":"(/client/[^"]+)"', response_text)
def main():
output_file = "mcpso_all_hrefs.json"
visited = set()
href_list = []
# Resume: load already saved hrefs
if os.path.exists(output_file):
with open(output_file, "r", encoding="utf-8") as f:
try:
href_list = json.load(f)
for item in href_list:
visited.add(item["href"])
except Exception:
pass
headers = {
"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36",
"Accept": "*/*",
"Accept-Encoding": "gzip, deflate, br, zstd",
"Accept-Language": "zh-CN,zh;q=0.9",
"Cookie": "Your Cookie",
"Next-Url": "/en/servers",
"Priority": "u=1, i",
"Referer": "https://mcp.so/servers",
"Rsc": "1",
"Sec-Ch-Ua": '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
"Sec-Ch-Ua-Arch": "arm",
"Sec-Ch-Ua-Bitness": "64",
"Sec-Ch-Ua-Full-Version": "136.0.7103.114",
"Sec-Ch-Ua-Full-Version-List": '"Chromium";v="136.0.7103.114", "Google Chrome";v="136.0.7103.114", "Not.A/Brand";v="99.0.0.0"',
"Sec-Ch-Ua-Mobile": "?0",
"Sec-Ch-Ua-Model": '""',
"Sec-Ch-Ua-Platform": '"macOS"',
"Sec-Ch-Ua-Platform-Version": '"15.3.0"',
"Sec-Fetch-Dest": "empty",
"Sec-Fetch-Mode": "cors",
"Sec-Fetch-Site": "same-origin"
}
for page in range(1, 477):
url = f"https://mcp.so/clients?page={page}"
print(f"Requesting page {page}: {url}")
for attempt in range(1, RETRY + 1):
try:
resp = requests.get(url, headers=headers, timeout=30)
if resp.status_code != 200:
print(f"Page {page} failed: HTTP {resp.status_code}, attempt {attempt}/{RETRY}")
time.sleep(SLEEP_BETWEEN_RETRY)
continue
hrefs = extract_hrefs(resp.text)
new_hrefs = [h for h in hrefs if h not in visited]
for h in new_hrefs:
href_list.append({"href": h})
visited.add(h)
# Save in real time
with open(output_file, "w", encoding="utf-8") as f:
json.dump(href_list, f, ensure_ascii=False, indent=2)
print(f"Page {page} got {len(new_hrefs)} new, total {len(href_list)}")
time.sleep(1)
break
except Exception as e:
print(f"Page {page} exception: {e}, attempt {attempt}/{RETRY}")
time.sleep(SLEEP_BETWEEN_RETRY)
continue
else:
print(f"Page {page} failed after {RETRY} retries.")
print(f"All done. Total unique hrefs: {len(href_list)}. Saved to {output_file}")
if __name__ == "__main__":
main()