--- license: apache-2.0 tags: - segmentation - mamba - wafer - electron-microscopy - tokenunify --- # TokenUnify Models This repository contains TokenUnify models of different sizes trained on wafer electron microscopy data, along with a superhuman baseline model. ## Available Models - **TokenUnify-1B.pth**: 1B parameter TokenUnify model - **TokenUnify-500M.pth**: 500M parameter TokenUnify model - **TokenUnify-200M.pth**: 200M parameter TokenUnify model - **TokenUnify-100M.pth**: 100M parameter TokenUnify model - **superhuman.pth**: Superhuman baseline model ## Model Details - **Architecture**: TokenUnify (based on Mamba) - **Training Data**: Wafer electron microscopy images - **Task**: Image Segmentation - **Framework**: PyTorch ## Usage ```python import torch # Load a specific model model_path = "TokenUnify-1B.pth" # or any other model file checkpoint = torch.load(model_path, map_location='cpu') # Your model loading code here ``` ## Model Sizes | Model | Parameters | File Name | |-------|------------|-----------| | TokenUnify Large | 1B | TokenUnify-1B.pth | | TokenUnify Medium | 500M | TokenUnify-500M.pth | | TokenUnify Small | 200M | TokenUnify-200M.pth | | TokenUnify Tiny | 100M | TokenUnify-100M.pth | | Superhuman Baseline | - | superhuman.pth | ## Citation If you use these models, please cite the relevant paper. EOF