generate-readme-eval / _script_for_gen.py
codelion's picture
Update _script_for_gen.py
99a5e82 verified
raw
history blame contribute delete
No virus
4.3 kB
import os
import subprocess
import json
import logging
from github import Github
import tiktoken
from pathlib import Path
import shutil
# Set up logging
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
# Initialize GitHub client
github_token = os.getenv('GITHUB_TOKEN')
if not github_token:
raise ValueError("GITHUB_TOKEN environment variable is not set")
g = Github(github_token)
# Initialize tokenizer
tokenizer = tiktoken.get_encoding("cl100k_base")
def clone_repo(repo_url, repo_name):
tmp_dir = f"/tmp/{repo_name}"
subprocess.run(["git", "clone", "--depth", "1", repo_url, tmp_dir], check=True)
return tmp_dir
def get_repo_content(repo_dir):
content = []
for root, dirs, files in os.walk(repo_dir):
if "test" in root.lower() or "example" in root.lower():
continue
for file in files:
if file.endswith('.py'):
file_path = os.path.join(root, file)
relative_path = os.path.relpath(file_path, repo_dir)
with open(file_path, 'r') as f:
file_content = f.read()
content.append(f"File: {relative_path}\n\n{file_content}\n\n")
return "\n".join(content)
def count_tokens(text):
return len(tokenizer.encode(text))
def process_repo(repo):
repo_name = repo.name
repo_url = repo.clone_url
logging.info(f"Processing repository: {repo_name}")
try:
tmp_dir = clone_repo(repo_url, repo_name)
readme_path = os.path.join(tmp_dir, "README.md")
if not os.path.exists(readme_path):
logging.info(f"README.md not found in {repo_name}")
return None
repo_content = get_repo_content(tmp_dir)
if count_tokens(repo_content) >= 100000:
logging.info(f"Repository {repo_name} content exceeds 100k tokens")
return None
with open(readme_path, 'r') as f:
readme_content = f.read()
repo_commit = subprocess.check_output(["git", "rev-parse", "HEAD"], cwd=tmp_dir).decode().strip()
return {
"repo_name": repo_name,
"repo_commit": repo_commit,
"repo_content": repo_content,
"repo_readme": readme_content
}
except Exception as e:
logging.error(f"Error processing repository {repo_name}: {str(e)}")
return None
finally:
if 'tmp_dir' in locals():
shutil.rmtree(tmp_dir)
def load_existing_data(filename):
existing_data = {}
if os.path.exists(filename):
with open(filename, "r") as f:
for line in f:
item = json.loads(line)
existing_data[item['repo_name']] = item
return existing_data
def save_dataset(filename, dataset, mode='a'):
with open(filename, mode) as f:
for item in dataset:
json.dump(item, f)
f.write("\n")
def main():
g = Github(os.getenv('GITHUB_TOKEN'))
filename = "generate-readme-eval.jsonl"
existing_data = load_existing_data(filename)
new_dataset = []
updated_count = 0
skipped_count = 0
repos = g.search_repositories(query="language:python stars:>1000 forks:>100", sort="stars", order="desc")
for i, repo in enumerate(repos[:400]):
if repo.full_name in existing_data:
existing_item = existing_data[repo.full_name]
if existing_item['repo_commit'] == repo.get_commits()[0].sha:
skipped_count += 1
logging.info(f"Skipped {repo.full_name}: Already processed with same commit")
continue
else:
logging.info(f"Updating {repo.full_name}: Commit changed")
updated_count += 1
item = process_repo(repo)
if item:
new_dataset.append(item)
if i % 10 == 0:
logging.info(f"Processed {i+1} repositories")
# Append new and updated items to the file
save_dataset(filename, new_dataset, mode='a')
logging.info(f"Dataset updated with {len(new_dataset)} new/updated items")
logging.info(f"Skipped {skipped_count} repositories (no changes)")
logging.info(f"Updated {updated_count} repositories")
if __name__ == "__main__":
main()