You signed in with another tab or window.Reload to refresh your session.You signed out in another tab or window.Reload to refresh your session.You switched accounts on another tab or window.Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+31-21Lines changed: 31 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -210,7 +210,7 @@ Finally, to pack the prediction file into the submission format, please modify t
210
210
python tools/submit_results.py
211
211
```
212
212
213
-
Then you can submit the resulting pkl file to the test server(to go live by the end of March)and wait for the lottery :)
213
+
Then you can submit the resulting pkl file to the test server and wait for the lottery :)
214
214
215
215
We also provide a sample script`tools/eval_script.py` for evaluating the submission file and you can check it by yourself to ensure your submitted file has the correct format.
216
216
@@ -224,45 +224,55 @@ Note that the performance is a little different from the results provided in the
Note: As mentioned in the paper, due to much more instances annotated with our new tools and pipelines, we concatenate several simple prompts as more complex ones to ensure those prompts to be more accurate without potential ambiguity. The above table is the benchmark without complex prompts using the initial version of visual grounding data.
256
244
257
245
We found such data is much less than the main part though, it can boost the multi-modal model's performance a lot. Meanwhile, whether to include these data in the validation set is not much important. We provide the updated benchmark as below and update a version of visual grounding data via emails to the community.
Because the occupancy prediction models are a little large, we save them via OpenXLab and do not provide direct download links here. To download these checkpoints on OpenXLab, please run the following commands:
# Then you can cd EmbodiedScan to get all the pretrained models
273
+
```
264
274
265
-
Please see the[paper](./assets/EmbodiedScan.pdf) for more details of ourtwo benchmarks, fundamental 3D perception and language-groundedbenchmarks. This dataset is still scaling up and the benchmark is being polished and extended. Please stay tuned for our recent updates.
275
+
Please see the[paper](./assets/EmbodiedScan.pdf) for more details of our benchmarks. This dataset is still scaling up and the benchmark is being polished and extended. Please stay tuned for our recent updates.