| | |
| | import argparse |
| | import os |
| | from functools import partial |
| | from multiprocessing import Pool |
| | from pathlib import Path |
| |
|
| | import numpy as np |
| | import pandas as pd |
| | import requests |
| | import utils |
| | from PIL import Image |
| | from tqdm import tqdm |
| |
|
| | METADATA_FILE = "published_images.csv" |
| | METADATA_URL = "https://raw.githubusercontent.com/NationalGalleryOfArt/opendata/refs/heads/main/data" |
| | IMG_URL = "https://api.nga.gov/iiif/%s/full/%s/0/default.jpg" |
| | METADATA_FOLDER = "metadata" |
| | EXTENSION = ".jpg" |
| |
|
| |
|
| | def download_metadata(annotation_folder): |
| | output_folder = annotation_folder / METADATA_FOLDER |
| | output_folder.mkdir(exist_ok=True) |
| | url = f"{METADATA_URL}/{METADATA_FILE}" |
| | print(url) |
| | response = requests.get(url) |
| | if response.status_code == 200: |
| | with open(output_folder / METADATA_FILE, "wb") as f: |
| | f.write(response.content) |
| |
|
| |
|
| | def download_url(row): |
| | if np.isnan(row.maxpixels) or ( |
| | row.maxpixels > row.width and row.maxpixels > row.height |
| | ): |
| | url = IMG_URL % (row.uuid, "full") |
| | else: |
| | url = IMG_URL % (row.uuid, f"!{row.maxpixels},{row.maxpixels}") |
| | return url |
| |
|
| |
|
| | def download_item(item, output_folder): |
| | uuid, url = item |
| | try: |
| | if (output_folder / f"{uuid}{EXTENSION}").exists(): |
| | print("skipping", uuid, "already downloaded") |
| | return |
| | response = requests.get(url) |
| | if response.status_code == 200: |
| | with open(output_folder / f"{uuid}{EXTENSION}", "wb") as f: |
| | f.write(response.content) |
| | except: |
| | print("errored", item) |
| | return |
| |
|
| |
|
| | def remove_non_compliant_image(item, output_folder): |
| | uuid, max_pixels = item |
| | if np.isnan(max_pixels): |
| | return |
| | if not (output_folder / f"{uuid}{EXTENSION}").exists(): |
| | return |
| | img = Image.open(output_folder / f"{uuid}{EXTENSION}") |
| | if img.width > max_pixels or img.height > max_pixels: |
| | os.remove(output_folder / f"{uuid}{EXTENSION}") |
| | return uuid |
| |
|
| |
|
| | def reshape_image(rel_path, filename_size_map, output_folder): |
| | w, h = filename_size_map[rel_path] |
| | path = output_folder / f"{rel_path}" |
| | img = Image.open(path) |
| | if img.width != w or img.height != h: |
| | new_size = (w, h) |
| | resized_img = img.resize(new_size) |
| | resized_img.save(path) |
| |
|
| |
|
| | def main(args, workers=20): |
| | raw_folder = Path(args.raw_images_folder) |
| | processed_folder = Path(args.processed_images_folder) |
| | utils.setup(raw_folder) |
| | utils.setup(processed_folder) |
| | uuids = utils.get_image_ids(args.annotation_file) |
| | filename_size_map = utils.get_filename_size_map(args.annotation_file) |
| | if not ((raw_folder / METADATA_FOLDER) / METADATA_FILE).exists(): |
| | download_metadata(raw_folder) |
| |
|
| | metadata = pd.read_csv((raw_folder / METADATA_FOLDER) / METADATA_FILE) |
| | metadata["download_url"] = metadata.apply(download_url, axis=1) |
| | available_uuids = list(uuids.intersection(set(metadata["uuid"].tolist()))) |
| | print(len(available_uuids), "available for download out of", len(uuids), "target") |
| | url_data = list( |
| | metadata.set_index("uuid") |
| | .loc[available_uuids] |
| | .to_dict()["download_url"] |
| | .items() |
| | ) |
| |
|
| | download_single = partial(download_item, output_folder=(processed_folder)) |
| |
|
| | print("Preparing to download", len(url_data), "items") |
| | with Pool(20) as p: |
| | for _ in tqdm(p.imap(download_single, url_data), total=len(url_data)): |
| | continue |
| | check_img_size = partial( |
| | remove_non_compliant_image, output_folder=(processed_folder) |
| | ) |
| | max_pixels_dict_all = metadata.set_index("uuid").to_dict()["maxpixels"] |
| | max_pixels_dict = {item[0]: max_pixels_dict_all[item[0]] for item in url_data} |
| | print("Checking all images within size constraints") |
| | non_compliant = set() |
| | with Pool(20) as p: |
| | for each in tqdm( |
| | p.imap(check_img_size, max_pixels_dict.items()), total=len(max_pixels_dict) |
| | ): |
| | if each is not None: |
| | non_compliant.add(each) |
| | print(len(non_compliant), "not compliant size, removed") |
| |
|
| | reshape_single = partial( |
| | reshape_image, |
| | filename_size_map=(filename_size_map), |
| | output_folder=(processed_folder), |
| | ) |
| | rel_paths = os.listdir(args.processed_images_folder) |
| | print("Preparing to reshape", len(rel_paths), "items") |
| | with Pool(20) as p: |
| | for _ in tqdm(p.imap(reshape_single, rel_paths), total=len(rel_paths)): |
| | continue |
| |
|
| |
|
| | if __name__ == "__main__": |
| | parser = argparse.ArgumentParser() |
| | parser.add_argument("--annotation_file", help="Path to annotation file") |
| | parser.add_argument("--raw_images_folder", help="Path to downloaded images") |
| | parser.add_argument("--processed_images_folder", help="Path to processed images") |
| | args = parser.parse_args() |
| | main(args) |
| |
|