- Notifications
You must be signed in to change notification settings - Fork67
A simulation platform for versatile Embodied AI research and developments.
License
InternRobotics/InternUtopia
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
- [2025-07] InternUtopia 2.2.0 is released!
- [2025-07] Our project has been renamed to InternUtopia.
- [2025-02] GRUtopia 2.0 released!
- [2024-07] We release thepaper and demos of GRUtopia.
- Gym compatible env implementation.
- Easy-to-use pythonic config system to use out-of-the-boxsensors,controllers,robots andtasks.
- Examples of drivingdiverse robots and the corresponding policies.
- Benchmark and baseline forsocial navigation andmobile manipulation task.
- Teleportation tools withMocap andApple VisionPro.
- Physically accurateinteractive object assets that are ready for simulation.
- ProceduralIndoor Scene Generation withGRScenes-100.
- 🏠 About
- 📚 Getting Started
- 🏙️ Assets
- 📦 Benchmark & Method
- 👥 Support
- 📝 TODO List
- 🔗 Citation
- 📄 License
- 👏 Acknowledgements
Recent works have been exploring the scaling laws in the field of Embodied AI. Given the prohibitive costs of collecting real-world data, we believe theSimulation-to-Real (Sim2Real) paradigm is a more feasible path for scaling the learning of embodied models.
We introduce projectInternUtopia (aka. 桃源 in Chinese), a general-purpose research platform for embodied AGI.It features several advancements:
- 🏙️GRScenes, the scene dataset, includes 100k interactive finely annotated scenes. GRScenes covers 89 diverse scene categories, facilitating deployment of general robots across different scenarios.
- 🧑🤝🧑GRResidents, a Large Language Model (LLM) driven Non-Player Character (NPC) system that enables social interaction, task generation, and task assignment, thus simulatingsocial scenarios for embodied AI applications.
- 🤖GRBench, a collection of embodied AI benchmarks for assessing various capabilities of solving embodied tasks.
We hope that this work can alleviate the scarcity of high-quality data in this field and provide a more comprehensive assessment of embodied AI research.
- Ubuntu 20.04, 22.04
- NVIDIA Omniverse Isaac Sim 4.5.0
- Ubuntu 20.04/22.04 Operating System
- NVIDIA GPU (RTX 2070 or higher)
- NVIDIA GPU Driver (recommended version 535.216.01+)
- Docker (Optional)
- NVIDIA Container Toolkit (Optional)
- Conda
- Python 3.10.16 (3.10.* should be ok)
We provide the installation guidehere. You can install locally or use docker and verify the installation easily.
We provide detaileddocs for the basic usage of different modules supported in InternUtopia. Welcome to try and post your suggestions!
Note
📝First of all youMUST complete theUser Agreement for GRScenes-100 Dataset Access.
Then you can choose to download all assets (~80GB) or a minimum set (~500MB) to examine installation by running the following script withInternUtopia installed:
$ python -m internutopia.download_assets
The default path to store the downloaded assets is${PATH/TO/INTERNUTOPIA/ROOT}/internutopia/assets. Users have two ways to configure the asset path:
- Spcecify a custom path during download using
python -m internutopia.download_assets. - Set it later by running
python -m internutopia.set_assets_pathand entering the preferred directory.
If you want to separately download theGRScenes-100 scene assets, you can manually download them fromOpenDataLab,ModelScope andHuggingFace. Please refer to theinstructions for scene usage.
If you want to separately download robots and policy weights, you can manually download therobot directory from fromOpenDataLab,ModelScope andHuggingFace and move it to the root of the asset path.
We preliminarily establish three benchmarks for evaluating the capabilities of embodied agents from different aspects:Object Loco-Navigation,Social Loco-Navigation, andLoco-Manipulation. Please refer to theInternNav andInternManip for running the benchmarks.
Join ourWeChat support group orDiscord for any help.
- Release the paper with demos.
- Release the platform with basic functions and demo scenes.
- Release 100 curated scenes.
- Polish APIs and related codes.
- Full release and further updates.
- Release the baseline methods and benchmark data.
- Support multiple episodes.
- Vectorized env and batch execution.
- Training framework.
If you find our work helpful, please cite:
@inproceedings{grutopia,title={GRUtopia: Dream General Robots in a City at Scale},author={Wang, Hanqing and Chen, Jiahe and Huang, Wensi and Ben, Qingwei and Wang, Tai and Mi, Boyu and Huang, Tao and Zhao, Siheng and Chen, Yilun and Yang, Sizhe and Cao, Peizhou and Yu, Wenye and Ye, Zichao and Li, Jialun and Long, Junfeng and Wang, ZiRui and Wang, Huiling and Zhao, Ying and Tu, Zhongying and Qiao, Yu and Lin, Dahua and Pang Jiangmiao},year={2024},booktitle={arXiv},}
InternUtopia's simulation platform isMIT licensed. The open-sourced GRScenes are under theCreative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
- OmniGibson: We refer to OmniGibson for designs of oracle actions.
- RSL_RL: We use
rsl_rllibrary to train the control policies for legged robots. - ReferIt3D: We refer to the Sr3D's approach to extract spatial relationship.
- Isaac Lab: We use some utilities from Orbit (Isaac Lab) for driving articulated joints in Isaac Sim.
- Open-TeleVision: We use Open-TeleVision to teleoperate with Apple VisionPro.
- HaMeR: We use HaMeR to recognize hand gesture in teleoperate with camera.
- Infinigen: We use Infinigen to procedurally generate indoor scenes uponGRScenes-100 dataset.
- VLFM: We refer to VLFM to implement our benchmark baselines.
- Grounding DINO: We use Grounding DINO in our benchmark baselines.
- YOLOv7: We use YOLOv7 in our benchmark baselines.
- MobileSAM: We use MobileSAM in our benchmark baselines.
About
A simulation platform for versatile Embodied AI research and developments.
Resources
License
Contributing
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Packages0
Uh oh!
There was an error while loading.Please reload this page.

