File size: 4,035 Bytes
9999996
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
import os, time, glob, datetime
from huggingface_hub import HfApi

TOKEN = "HF_TOKEN_REDACTED"
REPO = "OpenTransformer/web-crawl-2026"
STALE_SEC = 300  # 5 min no writes = done

api = HfApi(token=TOKEN)

def log(msg):
    ts = datetime.datetime.utcnow().isoformat()
    print(f"{ts} {msg}", flush=True)

def find_ready_files():
    ready = []
    now = time.time()
    for d in ["/workspace/staging", "/workspace/scraped_data_go", "/workspace/scraped_data_rust", "/workspace/scraped_data"]:
        for f in glob.glob(os.path.join(d, "*.gz")):
            age = now - os.path.getmtime(f)
            sz = os.path.getsize(f)
            if age > STALE_SEC and sz > 1024*1024:  # stale and >1MB
                ready.append((f, sz, age))
                log(f"  Ready: {f} ({sz/(1024*1024):.0f}MB, {age/3600:.1f}h old)")
    return ready

def upload_file(filepath, size):
    ts = datetime.datetime.utcnow().strftime("%Y%m%d_%H%M%S")
    basename = os.path.basename(filepath)
    remote = f"crawl/combined/{basename.replace(chr(46)+chr(106),chr(95)+ts+chr(46)+chr(106))}"
    log(f"Uploading {basename} ({size/(1024*1024):.0f}MB) -> {remote}")
    try:
        api.upload_file(
            path_or_fileobj=filepath,
            path_in_repo=remote,
            repo_id=REPO,
            repo_type="dataset",
            commit_message=f"Crawl data: {basename} ({size/(1024*1024):.0f}MB)"
        )
        log(f"Uploaded! Removing {filepath}")
        os.remove(filepath)
        return True
    except Exception as e:
        log(f"Upload failed: {e}")
        return False

def combine_and_upload(files):
    ts = datetime.datetime.utcnow().strftime("%Y%m%d_%H%M%S")
    combined = f"/workspace/crawl_batch_{ts}.jsonl.gz"
    total = sum(s for _, s, _ in files)
    log(f"Combining {len(files)} files ({total/(1024*1024):.0f}MB)")
    with open(combined, "wb") as out:
        for f, _, _ in files:
            with open(f, "rb") as inp:
                while True:
                    chunk = inp.read(8*1024*1024)
                    if not chunk:
                        break
                    out.write(chunk)
    remote = f"crawl/combined/crawl_batch_{ts}.jsonl.gz"
    final = os.path.getsize(combined)
    log(f"Uploading combined {final/(1024*1024):.0f}MB -> {remote}")
    try:
        api.upload_file(
            path_or_fileobj=combined,
            path_in_repo=remote,
            repo_id=REPO,
            repo_type="dataset",
            commit_message=f"Crawl batch {ts} ({final/(1024*1024):.0f}MB, {len(files)} files)"
        )
        log(f"Uploaded! Cleaning up...")
        for f, _, _ in files:
            os.remove(f)
        os.remove(combined)
        return True
    except Exception as e:
        log(f"Upload failed: {e}")
        if os.path.exists(combined):
            os.remove(combined)
        return False

def main():
    log("Upload daemon v2 starting")
    while True:
        log("Scanning...")
        ready = find_ready_files()
        if not ready:
            log("No files ready, sleeping 30min")
            time.sleep(1800)
            continue
        
        # If any single file >= 100MB, upload individually
        big = [(f, s, a) for f, s, a in ready if s >= 100*1024*1024]
        small = [(f, s, a) for f, s, a in ready if s < 100*1024*1024]
        
        for f, s, a in big:
            upload_file(f, s)
        
        # Combine small files if total >= 100MB, or if any are >12h old (avoid data loss)
        if small:
            total_small = sum(s for _, s, _ in small)
            max_age = max(a for _, _, a in small)
            if total_small >= 100*1024*1024 or max_age > 43200:  # 12 hours
                if len(small) == 1:
                    upload_file(small[0][0], small[0][1])
                else:
                    combine_and_upload(small)
            else:
                log(f"Small files total {total_small/(1024*1024):.0f}MB, newest {max_age/3600:.1f}h old, waiting")
        
        time.sleep(1800)

if __name__ == "__main__":
    main()