| | try: |
| | import detectron2 |
| | except: |
| | import os |
| | os.system('pip install git+https://github.com/facebookresearch/detectron2.git') |
| |
|
| | from inference import * |
| | import gradio as gr |
| | import glob |
| |
|
| | def gradio_app(image_path): |
| | """Helper function to run inference on provided image""" |
| |
|
| | predictions, out_pil = run_inference(image_path) |
| |
|
| | return out_pil |
| |
|
| |
|
| | |
| | |
| | |
| |
|
| | title = "MBARI Monterey Bay Benthic Supercategory" |
| | description = "Gradio demo for MBARI Monterey Bay Benthic Supercategory: This " \ |
| | "is a RetinaNet model fine-tuned from the Detectron2 object " \ |
| | "detection platform's ResNet backbone to identify 20 benthic " \ |
| | "supercategories drawn from MBARI's remotely operated vehicle " \ |
| | "image data collected in Monterey Bay off the coast of Central " \ |
| | "California. The data is drawn from FathomNet and consists of " \ |
| | "32779 images that contain a total of 80683 localizations. The " \ |
| | "model was trained on an 85/15 train/validation split at the " \ |
| | "image level. DOI: 10.5281/zenodo.5571043. " |
| |
|
| | examples = glob.glob("images/*.png") |
| |
|
| | interface = gr.Interface(gradio_app, |
| | inputs=[gr.components.Image(type="filepath")], |
| | outputs=gr.components.Image(type="pil"), |
| | title=title, |
| | description=description, |
| | examples=examples |
| | ) |
| |
|
| | interface.queue().launch() |
| |
|