| | import os |
| | import subprocess |
| | import json |
| | import logging |
| | from github import Github |
| | import tiktoken |
| | from pathlib import Path |
| | import shutil |
| |
|
| | |
| | logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s') |
| |
|
| | |
| | github_token = os.getenv('GITHUB_TOKEN') |
| | if not github_token: |
| | raise ValueError("GITHUB_TOKEN environment variable is not set") |
| | g = Github(github_token) |
| |
|
| | |
| | tokenizer = tiktoken.get_encoding("cl100k_base") |
| |
|
| | def clone_repo(repo_url, repo_name): |
| | tmp_dir = f"/tmp/{repo_name}" |
| | subprocess.run(["git", "clone", "--depth", "1", repo_url, tmp_dir], check=True) |
| | return tmp_dir |
| |
|
| | def get_repo_content(repo_dir): |
| | content = [] |
| | for root, dirs, files in os.walk(repo_dir): |
| | if "test" in root.lower() or "example" in root.lower(): |
| | continue |
| | for file in files: |
| | if file.endswith('.py'): |
| | file_path = os.path.join(root, file) |
| | relative_path = os.path.relpath(file_path, repo_dir) |
| | with open(file_path, 'r') as f: |
| | file_content = f.read() |
| | content.append(f"File: {relative_path}\n\n{file_content}\n\n") |
| | return "\n".join(content) |
| |
|
| | def count_tokens(text): |
| | return len(tokenizer.encode(text)) |
| |
|
| | def process_repo(repo): |
| | repo_name = repo.name |
| | repo_url = repo.clone_url |
| | logging.info(f"Processing repository: {repo_name}") |
| |
|
| | try: |
| | tmp_dir = clone_repo(repo_url, repo_name) |
| | readme_path = os.path.join(tmp_dir, "README.md") |
| | |
| | if not os.path.exists(readme_path): |
| | logging.info(f"README.md not found in {repo_name}") |
| | return None |
| |
|
| | repo_content = get_repo_content(tmp_dir) |
| | if count_tokens(repo_content) >= 100000: |
| | logging.info(f"Repository {repo_name} content exceeds 100k tokens") |
| | return None |
| |
|
| | with open(readme_path, 'r') as f: |
| | readme_content = f.read() |
| |
|
| | repo_commit = subprocess.check_output(["git", "rev-parse", "HEAD"], cwd=tmp_dir).decode().strip() |
| |
|
| | return { |
| | "repo_name": repo_name, |
| | "repo_commit": repo_commit, |
| | "repo_content": repo_content, |
| | "repo_readme": readme_content |
| | } |
| | except Exception as e: |
| | logging.error(f"Error processing repository {repo_name}: {str(e)}") |
| | return None |
| | finally: |
| | if 'tmp_dir' in locals(): |
| | shutil.rmtree(tmp_dir) |
| |
|
| | def load_existing_data(filename): |
| | existing_data = {} |
| | if os.path.exists(filename): |
| | with open(filename, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | existing_data[item['repo_name']] = item |
| | return existing_data |
| |
|
| | def save_dataset(filename, dataset, mode='a'): |
| | with open(filename, mode) as f: |
| | for item in dataset: |
| | json.dump(item, f) |
| | f.write("\n") |
| |
|
| | def main(): |
| | g = Github(os.getenv('GITHUB_TOKEN')) |
| | filename = "generate-readme-eval.jsonl" |
| | existing_data = load_existing_data(filename) |
| | |
| | new_dataset = [] |
| | updated_count = 0 |
| | skipped_count = 0 |
| |
|
| | repos = g.search_repositories(query="language:python stars:>1000 forks:>100", sort="stars", order="desc") |
| |
|
| | for i, repo in enumerate(repos[:400]): |
| | if repo.full_name in existing_data: |
| | existing_item = existing_data[repo.full_name] |
| | if existing_item['repo_commit'] == repo.get_commits()[0].sha: |
| | skipped_count += 1 |
| | logging.info(f"Skipped {repo.full_name}: Already processed with same commit") |
| | continue |
| | else: |
| | logging.info(f"Updating {repo.full_name}: Commit changed") |
| | updated_count += 1 |
| | |
| | item = process_repo(repo) |
| | if item: |
| | new_dataset.append(item) |
| | |
| | if i % 10 == 0: |
| | logging.info(f"Processed {i+1} repositories") |
| |
|
| | |
| | save_dataset(filename, new_dataset, mode='a') |
| |
|
| | logging.info(f"Dataset updated with {len(new_dataset)} new/updated items") |
| | logging.info(f"Skipped {skipped_count} repositories (no changes)") |
| | logging.info(f"Updated {updated_count} repositories") |
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|