File size: 6,358 Bytes
1fc5206
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
#!/usr/bin/env bash
# Runner script for Track 4: Decoder Topology experiments.
# Tests tree-aware decoding (teacher_forced / iterative) vs chain baseline.
# Outputs METRICS:{json} for Ratiocinator fleet parsing.
#
# Environment variables (set by Ratiocinator fleet):
#   DECODER_EDGE_MODE - chain, teacher_forced, iterative (default: teacher_forced)
#   DECODER_CONV_TYPE - GCN, SAGE, GAT, GIN, GraphConv (default: GAT)
#   HIDDEN_DIM        - Hidden dimension (default: 256)
#   NUM_LAYERS        - Number of decoder layers (default: 5)
#   LEARNING_RATE     - Learning rate (default: 0.001)
#   TYPE_WEIGHT       - Weight for node type loss (default: 2.0)
#   PARENT_WEIGHT     - Weight for parent prediction loss (default: 1.0)
#   LOSS_FN           - Loss function: improved, comprehensive, simple, original (default: improved)
#   EPOCHS            - Training epochs (default: 30)
#   DATASET_PATH      - Path to dataset dir (default: dataset/)

set -uo pipefail

DECODER_EDGE_MODE="${DECODER_EDGE_MODE:-teacher_forced}"
DECODER_CONV_TYPE="${DECODER_CONV_TYPE:-GAT}"
HIDDEN_DIM="${HIDDEN_DIM:-256}"
NUM_LAYERS="${NUM_LAYERS:-5}"
LEARNING_RATE="${LEARNING_RATE:-0.001}"
TYPE_WEIGHT="${TYPE_WEIGHT:-2.0}"
PARENT_WEIGHT="${PARENT_WEIGHT:-1.0}"
LOSS_FN="${LOSS_FN:-improved}"
EPOCHS="${EPOCHS:-30}"
DATASET_PATH="${DATASET_PATH:-dataset/}"
OUTPUT_PATH="models/experiment_topology_decoder.pt"
ENCODER_PATH="models/best_model.pt"

echo "=== Track 4: Decoder Topology Arm ==="
echo "EDGE_MODE=$DECODER_EDGE_MODE DECODER=$DECODER_CONV_TYPE HIDDEN=$HIDDEN_DIM LAYERS=$NUM_LAYERS"
echo "LR=$LEARNING_RATE TYPE_W=$TYPE_WEIGHT PARENT_W=$PARENT_WEIGHT LOSS=$LOSS_FN"

# Pull LFS files if they are pointers (e.g., after shallow clone)
if command -v git-lfs &>/dev/null || git lfs version &>/dev/null 2>&1; then
    echo "Pulling LFS files..."
    git lfs pull 2>&1 || echo "LFS pull returned non-zero (may be OK if files exist)"
elif [ -f "${DATASET_PATH}/validation.jsonl" ] && head -1 "${DATASET_PATH}/validation.jsonl" | grep -q "^version https://git-lfs"; then
    echo "ERROR: LFS pointer files detected but git-lfs not installed"
    exit 1
fi

# Ensure train/val split exists
if [ ! -f "${DATASET_PATH}/train.jsonl" ]; then
    echo "Creating train/val split..."
    python scripts/split_complexity_data.py \
        --input "${DATASET_PATH}/validation.jsonl" \
        --output-dir "${DATASET_PATH}"
fi

# Symlink validation.jsonl → val.jsonl for compatibility
if [ -f "${DATASET_PATH}/val.jsonl" ]; then
    ORIG_VAL="${DATASET_PATH}/validation.jsonl"
    if [ -f "$ORIG_VAL" ] && ! [ -L "$ORIG_VAL" ]; then
        mv "$ORIG_VAL" "${DATASET_PATH}/validation_full.jsonl"
    fi
    ln -sf val.jsonl "${DATASET_PATH}/validation.jsonl"
fi

# Need pre-trained encoder
if [ ! -f "$ENCODER_PATH" ]; then
    echo "Training encoder first..."
    python train.py --epochs 20 --output_path "$ENCODER_PATH" --dataset_path "$DATASET_PATH" --num_workers 0
fi

mkdir -p models

# Run autoencoder training with tree-aware decoder — stream output
TRAIN_LOG="/tmp/topo_train_$$.log"
python train_autoencoder.py \
    --dataset_path "$DATASET_PATH" \
    --epochs "$EPOCHS" \
    --output_path "$OUTPUT_PATH" \
    --encoder_weights_path "$ENCODER_PATH" \
    --hidden_dim "$HIDDEN_DIM" \
    --num_layers "$NUM_LAYERS" \
    --decoder_conv_type "$DECODER_CONV_TYPE" \
    --decoder_edge_mode "$DECODER_EDGE_MODE" \
    --learning_rate "$LEARNING_RATE" \
    --type_weight "$TYPE_WEIGHT" \
    --parent_weight "$PARENT_WEIGHT" \
    --loss_fn "$LOSS_FN" \
    2>&1 | tee "$TRAIN_LOG"

TRAIN_RC=${PIPESTATUS[0]}
if [ "$TRAIN_RC" -ne 0 ]; then
    echo "ERROR: train_autoencoder.py exited with code $TRAIN_RC"
    echo "METRICS:{\"error\": \"training_failed\", \"exit_code\": $TRAIN_RC}"
    exit 1
fi

BEST_VAL_LOSS=$(grep "Best validation loss" "$TRAIN_LOG" | grep -oP '[\d.]+' | tail -1)

# Run syntactic validity evaluation
python -c "
import sys, os, json, torch
sys.path.insert(0, os.path.join(os.path.dirname('.'), 'src'))
from models import ASTAutoencoder
from data_processing import create_data_loaders

device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
num_samples = 100
valid_count = 0
total = 0

try:
    model = ASTAutoencoder(
        encoder_input_dim=74,
        node_output_dim=74,
        hidden_dim=$HIDDEN_DIM,
        num_layers=$NUM_LAYERS,
        conv_type='SAGE',
        freeze_encoder=True,
        encoder_weights_path='$ENCODER_PATH',
        decoder_conv_type='$DECODER_CONV_TYPE',
        decoder_edge_mode='$DECODER_EDGE_MODE',
    ).to(device)

    checkpoint = torch.load('$OUTPUT_PATH', map_location=device, weights_only=False)
    model.decoder.load_state_dict(checkpoint['decoder_state_dict'])
    model.eval()

    # Load val data
    val_path = os.path.join('${DATASET_PATH}', 'val.jsonl')
    if not os.path.exists(val_path):
        val_path = os.path.join('${DATASET_PATH}', 'validation.jsonl')
    _, val_loader = create_data_loaders(val_path, val_path, batch_size=1, shuffle=False, num_workers=0)

    with torch.no_grad():
        for batch in val_loader:
            if total >= num_samples:
                break
            batch = batch.to(device)
            result = model(batch)
            recon = result['reconstruction']
            node_feats = recon.get('node_features') if isinstance(recon, dict) else None
            if node_feats is not None:
                pred_types = node_feats.argmax(dim=-1)
                unique_types = len(pred_types.unique())
                if unique_types > 2:
                    valid_count += 1
            total += 1

    validity_pct = (valid_count / total * 100) if total > 0 else 0.0
except Exception as e:
    validity_pct = 0.0
    total = num_samples
    print(f'Eval error: {e}', file=sys.stderr)

print('METRICS:' + json.dumps({
    'syntactic_validity_pct': round(validity_pct, 2),
    'val_loss': round(float('${BEST_VAL_LOSS:-0}'), 4),
    'samples_evaluated': total,
    'valid_samples': valid_count,
    'decoder_edge_mode': '$DECODER_EDGE_MODE',
    'decoder_conv_type': '$DECODER_CONV_TYPE',
    'hidden_dim': $HIDDEN_DIM,
    'num_layers': $NUM_LAYERS,
    'loss_fn': '$LOSS_FN',
    'type_weight': $TYPE_WEIGHT,
    'parent_weight': $PARENT_WEIGHT,
    'learning_rate': $LEARNING_RATE,
    'epochs': $EPOCHS,
}))
" 2>&1

rm -f "$TRAIN_LOG"