kin-zhang commited on
Commit
49e04ea
Β·
verified Β·
1 Parent(s): 820dd06

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -21
README.md CHANGED
@@ -5,18 +5,26 @@ language:
5
  ---
6
  # Scene Flow Models for Autonomous Driving Dataset
7
 
8
- <!-- Provide a quick summary of what the model is/does. -->
 
 
 
 
 
 
9
 
10
- Please check the usage on [KTH-RPL/DeFlow](https://github.com/KTH-RPL/DeFlow) or [KTH-RPL/SeFlow](https://github.com/KTH-RPL/SeFlow).
11
 
12
- - [DeFlow](https://github.com/KTH-RPL/DeFlow): Supervised learning scene flow, model included is trained on Argoverse 2
13
- - [SeFlow](https://github.com/KTH-RPL/SeFlow): **Self-Supervised** learning scene flow, model included is trained on Argoverse 2. Paper also reported Waymo result, the weight cannot be shared according to [Waymo Term](https://waymo.com/open/terms/). More detail discussion [issue 8](https://github.com/KTH-RPL/SeFlow/issues/8#issuecomment-2464224813).
 
 
 
 
 
14
 
15
  The files we included and all test result reports can be found [v2 leaderboard](https://github.com/KTH-RPL/DeFlow/discussions/6) and [v1 leaderboard](https://github.com/KTH-RPL/DeFlow/discussions/2).
16
- * [seflow_best.ckpt](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/seflow_best.ckpt): trained on Argoverse 2 sensor set with self-supervised way.
17
- * [deflow_best.ckpt](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/deflow_best.ckpt): trained on Argoverse 2 sensor set with ground truth using deflowLoss function.
18
- * [fastflow3d_best.ckpt](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/fastflow3d_best.ckpt): trained on Argoverse 2 sensor set with ground truth using ff3dLoss function.
19
- * ... more models on the way
20
  * [demo_data.zip](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/demo_data.zip): 613Mb, a mini-dataset for user to quickly run train/val code. Check usage in [this section](https://github.com/KTH-RPL/SeFlow?tab=readme-ov-file#1-run--train).
21
  * [waymo_map.tar.gz](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/waymo_map.tar.gz): to successfully process waymo data with ground segmentation included to unified h5 file. Check usage in [this README](https://github.com/KTH-RPL/SeFlow/blob/main/dataprocess/README.md#waymo-dataset).
22
 
@@ -25,30 +33,23 @@ The files we included and all test result reports can be found [v2 leaderboard](
25
  <!-- <br> -->
26
  You can try following methods in our code without any effort to make your own benchmark.
27
 
 
 
28
  - [x] [SeFlow](https://arxiv.org/abs/2407.01702) (Ours πŸš€): ECCV 2024
29
  - [x] [DeFlow](https://arxiv.org/abs/2401.16122) (Ours πŸš€): ICRA 2024
30
  - [x] [FastFlow3d](https://arxiv.org/abs/2103.01306): RA-L 2021
31
  - [x] [ZeroFlow](https://arxiv.org/abs/2305.10424): ICLR 2024, their pre-trained weight can covert into our format easily through [the script](https://github.com/KTH-RPL/SeFlow/tools/zerof2ours.py).
32
  - [ ] [NSFP](https://arxiv.org/abs/2111.01253): NeurIPS 2021, faster 3x than original version because of [our CUDA speed up](https://github.com/KTH-RPL/SeFlow/assets/cuda/README.md), same (slightly better) performance. Done coding, public after review.
33
  - [ ] [FastNSF](https://arxiv.org/abs/2304.09121): ICCV 2023. Done coding, public after review.
34
- <!-- - [ ] [Flow4D](https://arxiv.org/abs/2407.07995): 1st supervise network in the new leaderboard. Done coding, public after review. -->
35
  - [ ] ... more on the way
36
 
37
  </details>
38
 
39
- ## Citation
40
 
41
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
42
 
43
- ```
44
- @inproceedings{zhang2024deflow,
45
- author={Zhang, Qingwen and Yang, Yi and Fang, Heng and Geng, Ruoyu and Jensfelt, Patric},
46
- booktitle={2024 IEEE International Conference on Robotics and Automation (ICRA)},
47
- title={{DeFlow}: Decoder of Scene Flow Network in Autonomous Driving},
48
- year={2024},
49
- pages={2105-2111},
50
- doi={10.1109/ICRA57147.2024.10610278}
51
- }
52
  @inproceedings{zhang2024seflow,
53
  author={Zhang, Qingwen and Yang, Yi and Li, Peizheng and Andersson, Olov and Jensfelt, Patric},
54
  title={{SeFlow}: A Self-Supervised Scene Flow Method in Autonomous Driving},
@@ -58,4 +59,41 @@ You can try following methods in our code without any effort to make your own be
58
  organization={Springer},
59
  doi={10.1007/978-3-031-73232-4_20},
60
  }
61
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  ---
6
  # Scene Flow Models for Autonomous Driving Dataset
7
 
8
+ <p align="center">
9
+ <a href="https://github.com/KTH-RPL/OpenSceneFlow">
10
+ <picture>
11
+ <img alt="opensceneflow" src="https://github.com/KTH-RPL/OpenSceneFlow/blob/main/assets/docs/logo.png?raw=true" width="600">
12
+ </picture><br>
13
+ </a>
14
+ </p>
15
 
16
+ πŸ’ž If you find [*OpenSceneFlow*](https://github.com/KTH-RPL/OpenSceneFlow) useful to your research, please cite [**our works** πŸ“–](#cite-us) and give [a star 🌟](https://github.com/KTH-RPL/OpenSceneFlow) as encouragement. (ΰ©­ΛŠκ’³β€‹Λ‹)੭✧
17
 
18
+ OpenSceneFlow is a codebase for point cloud scene flow estimation.
19
+ Please check the usage on [KTH-RPL/OpenSceneFlow](https://github.com/KTH-RPL/OpenSceneFlow).
20
+
21
+ <!-- - [DeFlow](https://arxiv.org/abs/2401.16122): Supervised learning scene flow, model included is trained on Argoverse 2.
22
+ - [SeFlow](https://arxiv.org/abs/2407.01702): **Self-Supervised** learning scene flow, model included is trained on Argoverse 2. Paper also reported Waymo result, the weight cannot be shared according to [Waymo Term](https://waymo.com/open/terms/). More detail discussion [issue 8](https://github.com/KTH-RPL/SeFlow/issues/8#issuecomment-2464224813).
23
+ - [SSF](https://arxiv.org/abs/2501.17821): Supervised learning long-range scene flow, model included is trained on Argoverse 2.
24
+ - [Flow4D](https://ieeexplore.ieee.org/document/10887254): Supervised learning 4D network scene flow, model included is trained on Argoverse 2. -->
25
 
26
  The files we included and all test result reports can be found [v2 leaderboard](https://github.com/KTH-RPL/DeFlow/discussions/6) and [v1 leaderboard](https://github.com/KTH-RPL/DeFlow/discussions/2).
27
+ * [ModelName_best].ckpt: means the model evaluated in the public leaderboard page provided by authors or our retrained with the best parameters.
 
 
 
28
  * [demo_data.zip](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/demo_data.zip): 613Mb, a mini-dataset for user to quickly run train/val code. Check usage in [this section](https://github.com/KTH-RPL/SeFlow?tab=readme-ov-file#1-run--train).
29
  * [waymo_map.tar.gz](https://huggingface.co/kin-zhang/OpenSceneFlow/blob/main/waymo_map.tar.gz): to successfully process waymo data with ground segmentation included to unified h5 file. Check usage in [this README](https://github.com/KTH-RPL/SeFlow/blob/main/dataprocess/README.md#waymo-dataset).
30
 
 
33
  <!-- <br> -->
34
  You can try following methods in our code without any effort to make your own benchmark.
35
 
36
+ - [x] [SSF](https://arxiv.org/abs/2501.17821) (Ours πŸš€): ICRA 2025
37
+ - [x] [Flow4D](https://ieeexplore.ieee.org/document/10887254): RA-L 2025
38
  - [x] [SeFlow](https://arxiv.org/abs/2407.01702) (Ours πŸš€): ECCV 2024
39
  - [x] [DeFlow](https://arxiv.org/abs/2401.16122) (Ours πŸš€): ICRA 2024
40
  - [x] [FastFlow3d](https://arxiv.org/abs/2103.01306): RA-L 2021
41
  - [x] [ZeroFlow](https://arxiv.org/abs/2305.10424): ICLR 2024, their pre-trained weight can covert into our format easily through [the script](https://github.com/KTH-RPL/SeFlow/tools/zerof2ours.py).
42
  - [ ] [NSFP](https://arxiv.org/abs/2111.01253): NeurIPS 2021, faster 3x than original version because of [our CUDA speed up](https://github.com/KTH-RPL/SeFlow/assets/cuda/README.md), same (slightly better) performance. Done coding, public after review.
43
  - [ ] [FastNSF](https://arxiv.org/abs/2304.09121): ICCV 2023. Done coding, public after review.
 
44
  - [ ] ... more on the way
45
 
46
  </details>
47
 
48
+ ## Cite Us
49
 
50
+ *OpenSceneFlow* is designed by [Qingwen Zhang](https://kin-zhang.github.io/) from DeFlow and SeFlow project. If you find it useful, please cite our works:
51
 
52
+ ```bibtex
 
 
 
 
 
 
 
 
53
  @inproceedings{zhang2024seflow,
54
  author={Zhang, Qingwen and Yang, Yi and Li, Peizheng and Andersson, Olov and Jensfelt, Patric},
55
  title={{SeFlow}: A Self-Supervised Scene Flow Method in Autonomous Driving},
 
59
  organization={Springer},
60
  doi={10.1007/978-3-031-73232-4_20},
61
  }
62
+ @inproceedings{zhang2024deflow,
63
+ author={Zhang, Qingwen and Yang, Yi and Fang, Heng and Geng, Ruoyu and Jensfelt, Patric},
64
+ booktitle={2024 IEEE International Conference on Robotics and Automation (ICRA)},
65
+ title={{DeFlow}: Decoder of Scene Flow Network in Autonomous Driving},
66
+ year={2024},
67
+ pages={2105-2111},
68
+ doi={10.1109/ICRA57147.2024.10610278}
69
+ }
70
+ @article{zhang2025himu,
71
+ title={HiMo: High-Speed Objects Motion Compensation in Point Cloud},
72
+ author={Zhang, Qingwen and Khoche, Ajinkya and Yang, Yi and Ling, Li and Sina, Sharif Mansouri and Andersson, Olov and Jensfelt, Patric},
73
+ year={2025},
74
+ journal={arXiv preprint arXiv:2503.00803},
75
+ }
76
+ ```
77
+
78
+ And our excellent collaborators works as followings:
79
+
80
+ ```bibtex
81
+ @article{kim2025flow4d,
82
+ author={Kim, Jaeyeul and Woo, Jungwan and Shin, Ukcheol and Oh, Jean and Im, Sunghoon},
83
+ journal={IEEE Robotics and Automation Letters},
84
+ title={Flow4D: Leveraging 4D Voxel Network for LiDAR Scene Flow Estimation},
85
+ year={2025},
86
+ volume={10},
87
+ number={4},
88
+ pages={3462-3469},
89
+ doi={10.1109/LRA.2025.3542327}
90
+ }
91
+ @article{khoche2025ssf,
92
+ title={SSF: Sparse Long-Range Scene Flow for Autonomous Driving},
93
+ author={Khoche, Ajinkya and Zhang, Qingwen and Sanchez, Laura Pereira and Asefaw, Aron and Mansouri, Sina Sharif and Jensfelt, Patric},
94
+ journal={arXiv preprint arXiv:2501.17821},
95
+ year={2025}
96
+ }
97
+ ```
98
+
99
+ Feel free to contribute your method and add your bibtex here by pull request!