| | --- |
| | datasets: |
| | - Elsafty |
| | - Chula |
| | - DSE |
| | library_name: timm |
| | license: cc-by-nc-4.0 |
| | pipeline_tag: image-feature-extraction |
| | tags: |
| | - red-blood-cells |
| | - hematology |
| | - medical-imaging |
| | - vision-transformer |
| | - dino |
| | - dinov2 |
| | - feature-extraction |
| | - foundation-model |
| | model-index: |
| | - name: RedDino-large |
| | results: |
| | - task: |
| | type: image-classification |
| | name: RBC Shape Classification |
| | dataset: |
| | name: Elsafty |
| | type: Classification |
| | metrics: |
| | - type: Weighted F1 |
| | value: 88.5 |
| | - type: Balanced Accuracy |
| | value: 89.1 |
| | - type: Accuracy |
| | value: 88.4 |
| | - type: Weighted F1 |
| | value: 83.9 |
| | - type: Balanced Accuracy |
| | value: 79.0 |
| | - type: Accuracy |
| | value: 85.0 |
| | - type: Weighted F1 |
| | value: 86.6 |
| | - type: Balanced Accuracy |
| | value: 60.1 |
| | - type: Accuracy |
| | value: 86.6 |
| | --- |
| | |
| | # RedDino: A Foundation Model for Red Blood Cell Analysis |
| |
|
| | **RedDino** is a self-supervised Vision Transformer foundation model specifically designed for **red blood cell (RBC)** image analysis, as presented in the paper [RedDino: A foundation model for red blood cell analysis](https://arxiv.org/abs/2508.08180). |
| |
|
| | It leverages a tailored version of the **DINOv2** framework, trained on a meticulously curated dataset of **1.25 million RBC images** from diverse acquisition modalities and sources. This model excels at extracting robust, general-purpose features for downstream hematology tasks such as **shape classification**, **morphological subtype recognition**, and **batch-effect–robust analysis**. |
| |
|
| | Unlike general-purpose models pretrained on natural images, RedDino incorporates hematology-specific augmentations, architectural tweaks, and RBC-tailored data preprocessing, enabling **state-of-the-art performance** on multiple RBC benchmarks. |
| |
|
| | > 🧠 Developed by [Luca Zedda](https://orcid.org/0009-0001-8488-1612), [Andrea Loddo](https://orcid.org/0000-0002-6571-3816), [Cecilia Di Ruberto](https://orcid.org/0000-0003-4641-0307), and [Carsten Marr](https://orcid.org/0000-0003-2154-4552) |
| | > 🏥 University of Cagliari & Helmholtz Munich |
| | > 📄 Preprint: [arXiv:2508.08180](https://arxiv.org/abs/2508.08180) |
| | > 💻 Code: [https://github.com/Snarci/RedDino](https://github.com/Snarci/RedDino) |
| |
|
| | --- |
| |
|
| | ## Model Details |
| |
|
| | - **Architecture:** ViT-large, patch size 14 |
| | - **SSL framework:** DINOv2 (customized for RBC morphology) |
| | - **Pretraining dataset:** Curated RBC images from 18 datasets (multiple modalities and sources) |
| | - **Embedding size:** 1024 |
| | - **Intended use:** RBC morphology classification, feature extraction, batch-effect–robust analysis |
| | Notes: |
| | - RBC-specific training strategy including removal of KoLeo regularizer and Sinkhorn-Knopp centering. |
| | - Training on smear patches (not only single cells) to enhance cross-source generalization. |
| |
|
| | ## Example Usage |
| | ```python |
| | from PIL import Image |
| | from torchvision import transforms |
| | import timm |
| | import torch |
| | # Load model from Hugging Face Hub |
| | model = timm.create_model("hf_hub:Snarcy/RedDino-large", pretrained=True) |
| | model.eval() |
| | device = "cuda" if torch.cuda.is_available() else "cpu" |
| | model.to(device) |
| | # Load and preprocess image |
| | image = Image.open("path/to/rbc_image.jpg").convert("RGB") |
| | transform = transforms.Compose([ |
| | transforms.Resize((224, 224)), |
| | transforms.ToTensor(), |
| | transforms.Normalize(mean=[0.485, 0.456, 0.406], |
| | std=[0.229, 0.224, 0.225]), |
| | ]) |
| | input_tensor = transform(image).unsqueeze(0).to(device) |
| | # Extract features |
| | with torch.no_grad(): |
| | embedding = model(input_tensor) |
| | ``` |
| |
|
| | ## Model Variants |
| |
|
| | RedDino comes in three sizes to suit different computational requirements and performance needs: |
| |
|
| | | Model Variant | Embedding Size | Parameters | Usage | |
| | |---------------|----------------|------------|--------| |
| | | **RedDino-small** | 384 | 22M | `timm.create_model("hf_hub:Snarcy/RedDino-small", pretrained=True)` | |
| | | **RedDino-base** | 768 | 86M | `timm.create_model("hf_hub:Snarcy/RedDino-base", pretrained=True)` | |
| | | **RedDino-large** | 1024 | 304M | `timm.create_model("hf_hub:Snarcy/RedDino-large", pretrained=True)` | |
| |
|
| | Choose the variant that best fits your computational budget and performance requirements. Larger models generally provide richer feature representations at the cost of increased computational overhead. |
| |
|
| | --- |
| |
|
| | ## Benchmark Results |
| |
|
| | RedDino was benchmarked on major RBC classification datasets—including Elsafty, Chula, and DSE—outperforming state-of-the-art baselines such as ResNet50, DinoBloom, and DINOv2. |
| |
|
| | | Model | Dataset | Metric | Linear Probing (wF1) | 1-NN (wF1) | 20-NN (wF1) | |
| | |-------------------|-----------|-------------|----------------------|------------|-------------| |
| | | ResNet50 | Elsafty | Weighted F1 | 77.6 ± 8.1 | 64.3 ± 4.8 | 66.2 ± 4.9 | |
| | | DinoBloom-S | Elsafty | Weighted F1 | 83.2 ± 8.2 | 73.1 ± 5.1 | 76.5 ± 4.2 | |
| | | DINOv2 (small) | Elsafty | Weighted F1 | 82.1 ± 8.2 | 73.5 ± 4.8 | 77.2 ± 4.6 | |
| | | RedDino small | Elsafty | Weighted F1 | 86.0 ± 7.0 | 76.8 ± 4.9 | 80.0 ± 4.5 | |
| | | RedDino base | Elsafty | Weighted F1 | 88.1 ± 4.9 | 78.8 ± 3.6 | 82.6 ± 2.8 | |
| | | RedDino large | Elsafty | Weighted F1 | 88.5 ± 5.5 | 78.5 ± 4.6 | 81.6 ± 4.7 | |
| |
|
| | On Chula and DSE datasets, RedDino consistently surpassed all other models in feature quality (linear probing) with average improvements of 2–4% over prior approaches in key metrics. |
| |
|
| | --- |
| |
|
| | ## Highlights |
| |
|
| | - **Foundation model** for RBC analysis trained on the largest available multi-source RBC image set: 1.25M+ images, using advanced CellPose-based instance segmentation and patch extraction. |
| | - **DINOv2-based self-supervised learning** for label-efficient pretraining and robust, transferable features. |
| | - **Model architecture and key innovations**: |
| | - Patch-based training (224×224 px) shown to outperform single-cell training. |
| | - Novel data augmentation via Albumentations (32 pixel-level strategies). |
| | - Removal of the Koleo regularizer and adoption of Sinkhorn-Knopp centering for improved representation in RBC-specific domains. |
| | - Suite of models (small, base, large) covering 22M–304M parameters. |
| | - **Generalization**: Strong adaptation across varied protocols, microscopes, and imaging sites. Demonstrated resistance to batch effects and out-of-domain variance. |
| | - **Interpretability tools**: PCA/UMAP visualizations reveal clustering by phenotype and batch, distinguishing abnormal cells (e.g., malaria, echinocytes). |
| | - **Easy deployment**: Models and code are available on [GitHub](https://github.com/Snarci/RedDino) and [Hugging Face](https://huggingface.co/collections/Snarcy/reddino-689a13e29241d2e5690202fc). |
| |
|
| | --- |
| |
|
| | ## 📝 Citation |
| |
|
| | If you use this model, please cite the following paper: |
| |
|
| | **RedDino: A foundation model for red blood cell analysis** |
| | Luca Zedda, Andrea Loddo, Cecilia Di Ruberto, Carsten Marr — 2025 |
| | Preprint: arXiv:2508.08180. https://arxiv.org/abs/2508.08180 |
| |
|
| | ```bibtex |
| | @misc{zedda2025reddinofoundationmodelred, |
| | title={RedDino: A foundation model for red blood cell analysis}, |
| | author={Luca Zedda and Andrea Loddo and Cecilia Di Ruberto and Carsten Marr}, |
| | year={2025}, |
| | eprint={2508.08180}, |
| | archivePrefix={arXiv}, |
| | primaryClass={cs.CV}, |
| | url={https://arxiv.org/abs/2508.08180}, |
| | } |
| | ``` |
| |
|
| | --- |
| |
|
| | ## Summary |
| |
|
| | RedDino is the first family of foundation models tailored for comprehensive red blood cell image analysis, using large-scale self-supervised learning to set new performance benchmarks and generalization standards for computational hematology. Models and pretrained weights are available for research and practical deployment. |