| |
|
| | --- |
| | license: apache-2.0 |
| | tags: |
| | - segmentation |
| | - mamba |
| | - wafer |
| | - electron-microscopy |
| | - tokenunify |
| | --- |
| | |
| | # TokenUnify Models |
| |
|
| | This repository contains TokenUnify models of different sizes trained on wafer electron microscopy data, along with a superhuman baseline model. |
| |
|
| | ## Available Models |
| |
|
| | - **TokenUnify-1B.pth**: 1B parameter TokenUnify model |
| | - **TokenUnify-500M.pth**: 500M parameter TokenUnify model |
| | - **TokenUnify-200M.pth**: 200M parameter TokenUnify model |
| | - **TokenUnify-100M.pth**: 100M parameter TokenUnify model |
| | - **superhuman.pth**: Superhuman baseline model |
| |
|
| | ## Model Details |
| |
|
| | - **Architecture**: TokenUnify (based on Mamba) |
| | - **Training Data**: Wafer electron microscopy images |
| | - **Task**: Image Segmentation |
| | - **Framework**: PyTorch |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | import torch |
| | |
| | # Load a specific model |
| | model_path = "TokenUnify-1B.pth" # or any other model file |
| | checkpoint = torch.load(model_path, map_location='cpu') |
| | |
| | # Your model loading code here |
| | ``` |
| |
|
| | ## Model Sizes |
| |
|
| | | Model | Parameters | File Name | |
| | |-------|------------|-----------| |
| | | TokenUnify Large | 1B | TokenUnify-1B.pth | |
| | | TokenUnify Medium | 500M | TokenUnify-500M.pth | |
| | | TokenUnify Small | 200M | TokenUnify-200M.pth | |
| | | TokenUnify Tiny | 100M | TokenUnify-100M.pth | |
| | | Superhuman Baseline | - | superhuman.pth | |
| |
|
| | ## Citation |
| |
|
| | If you use these models, please cite the relevant paper. |
| | EOF |
| |
|