Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

[CVPR 2025] Source codes for the paper "3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning"

License

NotificationsYou must be signed in to change notification settings

UMass-Embodied-AGI/3D-Mem

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation


CVPR 2025

Yuncong Yang,Han Yang,Jiachen Zhou,Peihao Chen,Hongxin Zhang,Yilun Du,Chuang Gan

Paper PDFProject Page


This is the official repository of3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning.


News

  • [2025/03] Inference code for A-EQA and GOAT-Bench is released.
  • [2025/02] 3D-Mem is accepted to CVPR 2025!
  • [2024/12]Paper is on arXiv.

Installation

Set up the conda environment (Linux, Python 3.9):

conda create -n 3dmem python=3.9 -y&& conda activate 3dmempip install torch==2.0.1 torchvision==0.15.2 --index-url https://download.pytorch.org/whl/cu118conda install -c conda-forge -c aihabitat habitat-sim=0.2.5 headless faiss-cpu=1.7.4 -yconda install https://anaconda.org/pytorch3d/pytorch3d/0.7.4/download/linux-64/pytorch3d-0.7.4-py39_cu118_pyt201.tar.bz2 -ypip install omegaconf==2.3.0 open-clip-torch==2.26.1 ultralytics==8.2.31 supervision==0.21.0 opencv-python-headless==4.10.* \ scikit-learn==1.4 scikit-image==0.22 open3d==0.18.0 hipart==1.0.4 openai==1.35.3 httpx==0.27.2

Run Evaluation

1 - Preparations

Dataset

Please download the train and val split ofHM3D, and specifythe path incfg/eval_aeqa.yaml andcfg/eval_goatbench.yaml. For example, if your download path is/your_path/hm3d/ thatcontains/your_path/hm3d/train/ and/your_path/hm3d/val/, you can set thescene_data_path in the config files as/your_path/hm3d/.

The test questions of A-EQA and GOAT-Bench are provided in thedata/ folder. For A-EQA, we provide two subsets of different size:aeqa_questions-41.json andaeqa_questions-184.json, whereaeqa_questions-184.json is the official subset provided by OpenEQA andaeqa_questions-41.json is a smaller subset for quick evaluation.For GOAT-Bench, we include the completeval_unseen split in this repository.

OpenAI API Setup

Please set up the endpoint and API key for the OpenAI API insrc/const.py.

2 - Run Evaluation on A-EQA

First run the following script to generate the predictions for the A-EQA dataset:

python run_aeqa_evaluation.py -cf cfg/eval_aeqa.yaml

To split tasks, you can add--start_ratio and--end_ratio to specify the range of tasks to evaluate. For example,to evaluate the first half of the dataset, you can run:

python run_aeqa_evaluation.py -cf cfg/eval_aeqa.yaml --start_ratio 0.0 --end_ratio 0.5

After the scripts finish, the results from all splits will be automatically aggregated and saved.

To evaluate the predictions with the pipeline from OpenEQA, you can refer tolink

3 - Run Evaluation on GOAT-Bench

You can directly run the following script:

python run_goatbench_evaluation.py -cf cfg/eval_goatbench.yaml

The results will be saved and printed after the script finishes. You can also split the task similarly by adding--start_ratio and--end_ratio.Note that GOAT-Bench provides 10 explore episodes for each scene, and by default we only test the first episode due to the time and resource constraints.You can also specify the episode to evaluate for each scene by setting--split.

4 - Save Visualization

The default evaluation config will save visualization results including topdown maps, egocentric views, memory snapshots, and frontier snapshots at each step. Although saving visualization is very helpful, it may slows down the evaluation process. Please make save_visualization false if you would like to run large-scale evaluation.

Acknowledgement

The codebase is built uponOpenEQA,Explore-EQA, andConceptGraph.We thank the authors for their great work.

Citing 3D-Mem

@misc{yang20243dmem3dscenememory,      title={3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning},       author={Yuncong Yang and Han Yang and Jiachen Zhou and Peihao Chen and Hongxin Zhang and Yilun Du and Chuang Gan},      year={2024},      eprint={2411.17735},      archivePrefix={arXiv},      primaryClass={cs.CV},      url={https://arxiv.org/abs/2411.17735}, }

About

[CVPR 2025] Source codes for the paper "3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning"

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages


[8]ページ先頭

©2009-2025 Movatter.jp