Tavish9 commited on
Commit
ba28072
·
verified ·
1 Parent(s): 837be2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -3
README.md CHANGED
@@ -1,3 +1,34 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ base_model:
6
+ - IPEC-COMMUNITY/spatialvla-4b-224-pt
7
+ pipeline_tag: image-text-to-text
8
+ library_name: transformers
9
+ ---
10
+
11
+ # SpatialVLA Fine-Tuned on fractal & bridge
12
+
13
+ This model was produced by fine-tuning the [SpatialVLA model](IPEC-COMMUNITY/spatialvla-4b-224-pt) via LoRA (r=32) on the fractal and bridge dataset. We made a few modifications to the training dataset to improve final performance (see the
14
+ [SpatialVLA paper](https://arxiv.org/abs/2501.15830) for details).
15
+
16
+ ## Usage Instructions
17
+
18
+ See the [SpatialVLA GitHub README](https://github.com/SpatialVLA/SpatialVLA/blob/main/README.md) for instructions on how to run and evaluate this model on WidowX Robot tasks.
19
+
20
+ ## Citation
21
+
22
+ **BibTeX:**
23
+
24
+ ```BibTeX
25
+ @misc{qu2025spatialvlaexploringspatialrepresentations,
26
+ title={SpatialVLA: Exploring Spatial Representations for Visual-Language-Action Model},
27
+ author={Delin Qu and Haoming Song and Qizhi Chen and Yuanqi Yao and Xinyi Ye and Yan Ding and Zhigang Wang and JiaYuan Gu and Bin Zhao and Dong Wang and Xuelong Li},
28
+ year={2025},
29
+ eprint={2501.15830},
30
+ archivePrefix={arXiv},
31
+ primaryClass={cs.RO},
32
+ url={https://arxiv.org/abs/2501.15830},
33
+ }
34
+ ```