| --- |
| base_model: |
| - declare-lab/nora-long |
| datasets: |
| - TomNickson/OpenX-Embodiment |
| - jxu124/OpenX-Embodiment |
| language: |
| - en |
| license: mit |
|
|
| pipeline_tag: robotics |
| --- |
| # NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-based Preference Rewards |
|
|
| [](https://declare-lab.github.io/nora-1.5) |
| [](https://huggingface.co/declare-lab/nora-1.5) |
| [](https://arxiv.org/abs/2511.14659) |
| [](https://github.com/declare-lab/nora-1.5) |
|  |
|
|
| π₯ Project NORA is supported by Gemini and Lambda Labs! We are thankful to them. |
|
|
| NORA-1.5 is a **Vision-Language-Action (VLA)** model that improves generalization and real-world decision making through **post-training with world-model-based and action-based preference rewards**. |
| The model builds upon the NORA foundation to achieve stronger **instruction following**, **closed-loop control**, and **real-robot success**, demonstrating reliability across **LIBERO** and **SimplerEnv** environments. |
|
|
| This repository consolidates the full open-source release of **model checkpoints**, **inference code**, **training code**, and **evaluation tools**, along with documentation and examples. |
|
|
| <p align="center"> |
| <img src="https://declare-lab.github.io/assets/images/nora-1.5-arxiv-teaser.png" width="100%"> |
| </p> |
|
|
|
|
| --- |
|
|
| ## π Project Website |
|
|
| π **https://declare-lab.github.io/nora-1.5** |
| |
| --- |
|
|
| ## π Key Features |
|
|
| - **Vision-Language-Action architecture** with enhanced **task completion rate** and **distraction rate** |
| - **Action-based preference optimization** using expert preference rewards |
| - **World-model-based preference learning** for improved planning and consistency |
| - Strong **closed-loop control**, enabling deployment in real robot settings |
| - Supports **multi-task**, **long-horizon**, and **few-shot generalization** |
| - Compatible with **LeRobot**, **LIBERO**, **SimplerEnv**, and custom environments |
|
|
| --- |
|
|
| ## π¦ Repository Structure (will update) |
|
|
|
|
|
|
| ## π TODO <a name="todos"></a> ~ 1 week |
| - [ ] Release the inference code of Nora-1.5 |
| - [ ] Release all relevant model checkpoints(Pretrained, libero, SimplerEnv etc) |
| - [ ] Release the training/fine-tuning code of Nora-1.5 with LeRobot Dataset |
| - [ ] Release SimplerEnv evaluation code |
|
|
| ## Minimal Inference Sample (Will update) |
| ```python |
| from inference.modelling_expert import VLAWithExpert |
| |
| model = VLAWithExpert() |
| model.to('cuda') |
| outputs = model.sample_actions(PIL IMAGE,instruction,num_steps=10) ## Outputs 7 Dof action of normalized and unnormalized action |
| ``` |
|
|
| ## Citation |
|
|
| ```bibtex |
| @article{hung2025nora15, |
| title={NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-Based Preference Rewards}, |
| author={Hung, Chia-Yu and Majumder, Navonil and Deng, Haoyuan, Liu Renhang, Yankang Ang, Amir Zadeh, Chuan Li, Dorien Herremans, Ziwei Wang, and Soujanya Poria}, |
| journal={arXiv preprint}, |
| year={2025} |
| } |
| ``` |