Spaces:
Running
Running
Commit
Β·
c4e40c4
1
Parent(s):
fd54515
- Update docs
Browse files
README.md
CHANGED
@@ -38,7 +38,7 @@ pip install -r requirements.txt
|
|
38 |
|
39 |
### Download trained weights
|
40 |
|
41 |
-
- Download model weights and put it in the folder `weights`. You may also need to download the weights of [DPT model]() (a rgb2depth model). The `weights` folder will look like this:
|
42 |
|
43 |
```bash
|
44 |
βββ weights
|
@@ -55,6 +55,9 @@ pip install -r requirements.txt
|
|
55 |
streamlit run streamlit_apps/app.py --server.port 9113 --browser.gatherUsageStats False --server.fileWatcherType none
|
56 |
```
|
57 |
|
|
|
|
|
|
|
58 |
## Datasets
|
59 |
|
60 |
### COME15K dataset
|
@@ -74,6 +77,19 @@ streamlit run streamlit_apps/app.py --server.port 9113 --browser.gatherUsageStat
|
|
74 |
}
|
75 |
```
|
76 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
77 |
## References
|
78 |
|
79 |
All references are cited in these files:
|
|
|
38 |
|
39 |
### Download trained weights
|
40 |
|
41 |
+
- Download model weights and put it in the folder `weights`. You may also need to download the weights of [DPT model](https://drive.google.com/file/d/1vU4G31_T2PJv1DkA8j-MLXfMjGa7kD3L/view?usp=sharing) (a rgb2depth model). The `weights` folder will look like this:
|
42 |
|
43 |
```bash
|
44 |
βββ weights
|
|
|
55 |
streamlit run streamlit_apps/app.py --server.port 9113 --browser.gatherUsageStats False --server.fileWatcherType none
|
56 |
```
|
57 |
|
58 |
+

|
59 |
+

|
60 |
+
|
61 |
## Datasets
|
62 |
|
63 |
### COME15K dataset
|
|
|
77 |
}
|
78 |
```
|
79 |
|
80 |
+
## Acknowledgements
|
81 |
+
|
82 |
+
S-MultiMAE is build on top of [MultiMAE](https://github.com/EPFL-VILAB/MultiMAE). We kindly thank the authors for releasing their code.
|
83 |
+
|
84 |
+
```bib
|
85 |
+
@article{bachmann2022multimae,
|
86 |
+
author = {Roman Bachmann and David Mizrahi and Andrei Atanov and Amir Zamir},
|
87 |
+
title = {{MultiMAE}: Multi-modal Multi-task Masked Autoencoders},
|
88 |
+
booktitle = {European Conference on Computer Vision},
|
89 |
+
year = {2022},
|
90 |
+
}
|
91 |
+
```
|
92 |
+
|
93 |
## References
|
94 |
|
95 |
All references are cited in these files:
|
docs/streamlit_samples/sample1_input.png
ADDED
![]() |
docs/streamlit_samples/sample1_results.png
ADDED
![]() |
streamlit_apps/app_utils/image_inference.py
CHANGED
@@ -60,7 +60,9 @@ def image_inference(
|
|
60 |
disabled=img_file_buffer is None,
|
61 |
)
|
62 |
if is_predict:
|
63 |
-
with st.spinner(
|
|
|
|
|
64 |
start_time = time.time()
|
65 |
pred_depth, pred_sods, pred_sms = base_inference(
|
66 |
depth_model,
|
|
|
60 |
disabled=img_file_buffer is None,
|
61 |
)
|
62 |
if is_predict:
|
63 |
+
with st.spinner(
|
64 |
+
"Processing... (It usually takes about 30s - 1 minute per a set of salient objects)"
|
65 |
+
):
|
66 |
start_time = time.time()
|
67 |
pred_depth, pred_sods, pred_sms = base_inference(
|
68 |
depth_model,
|