| |
|
|
| """ |
| Script to extract and analyze training results from Roboflow VL100 experiments. |
| |
| This script processes training logs and configuration files to extract model performance |
| metrics and training parameters for analysis and comparison. |
| """ |
|
|
| import argparse |
| import json |
| import os |
| from typing import Any, Dict, List, Optional |
|
|
| import pandas as pd |
| import yaml |
|
|
|
|
| |
| CONFIG_FILENAME = "config_resolved.yaml" |
| RESULTS_FILENAME = "val_stats.json" |
| BBOX_AP_METRIC = "Meters_train/val_roboflow100/detection/coco_eval_bbox_AP" |
|
|
| |
| ROBOFLOW_CATEGORIES = { |
| "sports": [ |
| "actions", |
| "aerial-pool", |
| "ball", |
| "bibdetection", |
| "football-player-detection", |
| "lacrosse-object-detection", |
| ], |
| "other": [ |
| "buoy-onboarding", |
| "car-logo-detection", |
| "clashroyalechardetector", |
| "cod-mw-warzone", |
| "countingpills", |
| "everdaynew", |
| "flir-camera-objects", |
| "halo-infinite-angel-videogame", |
| "mahjong", |
| "new-defects-in-wood", |
| "orionproducts", |
| "pill", |
| "soda-bottles", |
| "taco-trash-annotations-in-context", |
| "the-dreidel-project", |
| ], |
| "aerial": [ |
| "aerial-airport", |
| "aerial-cows", |
| "aerial-sheep", |
| "apoce-aerial-photographs-for-object-detection-of-construction-equipment", |
| "electric-pylon-detection-in-rsi", |
| "floating-waste", |
| "human-detection-in-floods", |
| "sssod", |
| "uavdet-small", |
| "wildfire-smoke", |
| "zebrasatasturias", |
| ], |
| "medical": [ |
| "canalstenosis", |
| "crystal-clean-brain-tumors-mri-dataset", |
| "dentalai", |
| "inbreast", |
| "liver-disease", |
| "nih-xray", |
| "spinefrxnormalvindr", |
| "stomata-cells", |
| "train", |
| "ufba-425", |
| "urine-analysis1", |
| "x-ray-id", |
| "xray", |
| ], |
| "document": [ |
| "activity-diagrams", |
| "all-elements", |
| "circuit-voltages", |
| "invoice-processing", |
| "label-printing-defect-version-2", |
| "macro-segmentation", |
| "paper-parts", |
| "signatures", |
| "speech-bubbles-detection", |
| "wine-labels", |
| ], |
| "industrial": [ |
| "-grccs", |
| "13-lkc01", |
| "2024-frc", |
| "aircraft-turnaround-dataset", |
| "asphaltdistressdetection", |
| "cable-damage", |
| "conveyor-t-shirts", |
| "dataconvert", |
| "deeppcb", |
| "defect-detection", |
| "fruitjes", |
| "infraredimageofpowerequipment", |
| "ism-band-packet-detection", |
| "l10ul502", |
| "needle-base-tip-min-max", |
| "recode-waste", |
| "screwdetectclassification", |
| "smd-components", |
| "truck-movement", |
| "tube", |
| "water-meter", |
| "wheel-defect-detection", |
| ], |
| "flora_fauna": [ |
| "aquarium-combined", |
| "bees", |
| "deepfruits", |
| "exploratorium-daphnia", |
| "grapes-5", |
| "grass-weeds", |
| "gwhd2021", |
| "into-the-vale", |
| "jellyfish", |
| "marine-sharks", |
| "orgharvest", |
| "peixos-fish", |
| "penguin-finder-seg", |
| "pig-detection", |
| "roboflow-trained-dataset", |
| "sea-cucumbers-new-tiles", |
| "thermal-cheetah", |
| "tomatoes-2", |
| "trail-camera", |
| "underwater-objects", |
| "varroa-mites-detection--test-set", |
| "wb-prova", |
| "weeds4", |
| ], |
| } |
|
|
|
|
| def load_jsonl_last_row(file_path: str, keys: List[str]) -> Optional[Dict[str, Any]]: |
| """ |
| Load the last row from a JSONL file and extract specific keys. |
| |
| Args: |
| file_path: Path to the JSONL file |
| keys: List of keys to extract from the last row |
| |
| Returns: |
| Dictionary with extracted key-value pairs, or None if file not found/empty |
| """ |
| if not os.path.exists(file_path): |
| print(f"Warning: File not found: {file_path}") |
| return None |
|
|
| last_row = None |
| try: |
| with open(file_path, "r") as file: |
| for line in file: |
| last_row = json.loads(line.strip()) |
|
|
| if last_row is None: |
| print(f"Warning: Empty JSONL file: {file_path}") |
| return None |
|
|
| return {key: last_row.get(key) for key in keys} |
|
|
| except json.JSONDecodeError as e: |
| print(f"Error: Failed to parse JSON in {file_path}: {e}") |
| return None |
| except Exception as e: |
| print(f"Error: Failed to read {file_path}: {e}") |
| return None |
|
|
|
|
| def find_config_files(directory: str, filename: str = CONFIG_FILENAME) -> List[str]: |
| """ |
| Recursively find configuration files with a specific filename. |
| |
| Args: |
| directory: Root directory to search |
| filename: Target filename to search for |
| |
| Returns: |
| List of full paths to matching files |
| """ |
| matching_files = [] |
| for root, _, files in os.walk(directory): |
| |
| if "/code/" in root: |
| continue |
| if filename in files: |
| matching_files.append(os.path.join(root, filename)) |
| return matching_files |
|
|
|
|
| def extract_config_parameters(config_path: str, keys: List[str]) -> Dict[str, Any]: |
| """ |
| Extract specific parameters from a YAML configuration file. |
| |
| Args: |
| config_path: Path to the YAML configuration file |
| keys: List of keys to extract from the 'scratch' section |
| |
| Returns: |
| Dictionary containing extracted parameters |
| """ |
| try: |
| with open(config_path, "r") as file: |
| data = yaml.safe_load(file) |
|
|
| |
| scratch_params = {key: data["scratch"].get(key) for key in keys} |
|
|
| |
| launcher = data.get("launcher", {}) |
| scratch_params["batch_size"] = int(launcher.get("gpus_per_node", 1)) * int( |
| launcher.get("num_nodes", 1) |
| ) |
| scratch_params["lr_scale"] = data["scratch"].get("lr_scale") |
|
|
| roboflow_train = data.get("roboflow_train", {}) |
| scratch_params["roboflow_num_images"] = roboflow_train.get("num_images") |
|
|
| return scratch_params |
|
|
| except Exception as e: |
| print(f"Error: Failed to parse config file {config_path}: {e}") |
| return {} |
|
|
|
|
| def calculate_average(values_dict: Dict[str, float]) -> float: |
| """ |
| Calculate the average of values in a dictionary. |
| |
| Args: |
| values_dict: Dictionary with numeric values |
| |
| Returns: |
| Average of all values, or 0 if empty |
| """ |
| if not values_dict: |
| return 0.0 |
| return sum(values_dict.values()) / len(values_dict) |
|
|
|
|
| def extract_category_results(log_dir: str, categories: List[str]) -> Dict[str, float]: |
| """ |
| Extract bbox AP results for specific categories from log files. |
| |
| Args: |
| log_dir: Directory containing category log subdirectories |
| categories: List of category names to extract results for |
| |
| Returns: |
| Dictionary mapping category names to bbox AP scores |
| """ |
| results = {} |
| metric_keys = [BBOX_AP_METRIC] |
|
|
| for category in categories: |
| result_file = os.path.join(log_dir, f"logs/{category}/{RESULTS_FILENAME}") |
| category_result = load_jsonl_last_row(result_file, metric_keys) |
|
|
| if category_result is not None and category_result[BBOX_AP_METRIC] is not None: |
| results[category] = category_result[BBOX_AP_METRIC] |
|
|
| return results |
|
|
|
|
| def analyze_experiment_results(config_path: str) -> None: |
| """ |
| Analyze results from a single experiment configuration. |
| |
| Args: |
| config_path: Path to the experiment configuration file |
| """ |
| print("=" * 80) |
| print(f"Analyzing experiment: {config_path}") |
| print("=" * 80) |
|
|
| |
| config_keys = [ |
| "lr_transformer", |
| "lr_vision_backbone", |
| "lr_language_backbone", |
| "max_data_epochs", |
| ] |
|
|
| config_params = extract_config_parameters(config_path, config_keys) |
| print("Configuration Parameters:") |
| for key, value in config_params.items(): |
| print(f" {key}: {value}") |
| print() |
|
|
| |
| experiment_dir = os.path.dirname(config_path) |
| category_results = {} |
| category_averages = {} |
| all_scores = [] |
|
|
| for super_category, categories in ROBOFLOW_CATEGORIES.items(): |
| category_results[super_category] = extract_category_results( |
| experiment_dir, categories |
| ) |
|
|
| if category_results[super_category]: |
| category_averages[super_category] = calculate_average( |
| category_results[super_category] |
| ) |
| all_scores.extend(category_results[super_category].values()) |
|
|
| |
| print("Results by Category:") |
| for super_category, avg_score in category_averages.items(): |
| num_categories = len(category_results[super_category]) |
| print(f" {super_category}: {avg_score:.4f} (n={num_categories})") |
|
|
| print(f"\nOverall Results:") |
| print(f" Weighted average: {calculate_average(category_averages):.4f}") |
| print(f" Total categories: {len(all_scores)}") |
| print(f" True average: {sum(all_scores) / len(all_scores):.4f}") |
| print() |
|
|
|
|
| def print_results_table(results_data: List[Dict[str, Any]]) -> None: |
| """ |
| Print results in a formatted table. |
| |
| Args: |
| results_data: List of dictionaries containing results data |
| """ |
| if not results_data: |
| print("No results data to display.") |
| return |
|
|
| df = pd.DataFrame(results_data) |
| print("\nResults Summary Table:") |
| print("=" * 60) |
| print(df.to_string(index=False)) |
|
|
|
|
| def main() -> None: |
| """Main function to orchestrate the results extraction and analysis.""" |
| parser = argparse.ArgumentParser( |
| description="Extract and analyze Roboflow VL100 training results" |
| ) |
| parser.add_argument( |
| "-p", |
| "--path", |
| type=str, |
| required=True, |
| help="Root directory path containing experiment results", |
| ) |
|
|
| args = parser.parse_args() |
|
|
| |
| config_files = find_config_files(args.path, CONFIG_FILENAME) |
|
|
| if not config_files: |
| print(f"No configuration files found in {args.path}") |
| return |
|
|
| print(f"Found {len(config_files)} experiment configurations") |
| print() |
|
|
| |
| for config_file in config_files: |
| try: |
| analyze_experiment_results(config_file) |
| except Exception as e: |
| print(f"Error analyzing {config_file}: {e}") |
| continue |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|