from huggingface_hub import HfApi import tempfile, os api = HfApi(token='HF_TOKEN_REDACTED') REPO = 'OpenTransformer/web-crawl-2026' # Upload main.rs print('Uploading main.rs...') api.upload_file( path_or_fileobj='/workspace/rust_crawler/src/main.rs', path_in_repo='crawler/rust/src/main.rs', repo_id=REPO, repo_type='dataset', commit_message='Add Rust web crawler source (v3, 150-300 docs/s)' ) # Upload Cargo.toml print('Uploading Cargo.toml...') api.upload_file( path_or_fileobj='/workspace/rust_crawler/Cargo.toml', path_in_repo='crawler/rust/Cargo.toml', repo_id=REPO, repo_type='dataset', commit_message='Add Rust crawler Cargo.toml' ) # Create and upload README readme = '''--- license: apache-2.0 task_categories: - text-generation language: - en tags: - web-crawl - pretraining - nlp - text-corpus pretty_name: Web Crawl 2026 size_categories: - 10B