WXinlong
commited on
Commit
•
42fbb33
1
Parent(s):
d137d5b
first init
Browse files- README.md +50 -0
- seggpt_teaser.png +0 -0
README.md
CHANGED
@@ -1,3 +1,53 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
|
5 |
+
<div align="center">
|
6 |
+
<h1>SegGPT: Segmenting Everything In Context </h1>
|
7 |
+
|
8 |
+
[Xinlong Wang](https://www.xloong.wang/)<sup>1*</sup>, [Xiaosong Zhang](https://scholar.google.com/citations?user=98exn6wAAAAJ&hl=en)<sup>1*</sup>, [Yue Cao](http://yue-cao.me/)<sup>1*</sup>, [Wen Wang](https://scholar.google.com/citations?user=1ks0R04AAAAJ&hl)<sup>2</sup>, [Chunhua Shen](https://cshen.github.io/)<sup>2</sup>, [Tiejun Huang](https://scholar.google.com/citations?user=knvEK4AAAAAJ&hl=en)<sup>1,3</sup>
|
9 |
+
|
10 |
+
<sup>1</sup>[BAAI](https://www.baai.ac.cn/english.html), <sup>2</sup>[ZJU](https://www.zju.edu.cn/english/), <sup>3</sup>[PKU](https://english.pku.edu.cn/)
|
11 |
+
|
12 |
+
Enjoy the [Demo](https://huggingface.co/spaces/BAAI/SegGPT) and [Code](https://github.com/baaivision/Painter/edit/main/SegGPT)
|
13 |
+
|
14 |
+
|
15 |
+
<br>
|
16 |
+
|
17 |
+
<image src="seggpt_teaser.png" width="720px" />
|
18 |
+
<br>
|
19 |
+
|
20 |
+
</div>
|
21 |
+
|
22 |
+
<br>
|
23 |
+
|
24 |
+
We present SegGPT, a generalist model for segmenting everything in context. With only one single model, SegGPT can perform arbitrary segmentation tasks in images or videos via in-context inference, such as object instance, stuff, part, contour, and text.
|
25 |
+
SegGPT is evaluated on a broad range of tasks, including few-shot semantic segmentation, video object segmentation, semantic segmentation, and panoptic segmentation.
|
26 |
+
Our results show strong capabilities in segmenting in-domain and out-of-domain targets, either qualitatively or quantitatively.
|
27 |
+
|
28 |
+
[[Paper]](https://arxiv.org/abs/2304.03284)
|
29 |
+
[[Code]](https://github.com/baaivision/Painter/edit/main/SegGPT)
|
30 |
+
[[Demo]](https://huggingface.co/spaces/BAAI/SegGPT)
|
31 |
+
|
32 |
+
## **Model**
|
33 |
+
|
34 |
+
A pre-trained SegGPT model is available at [🤗 HF link](https://huggingface.co/BAAI/Painter/blob/main/painter_vit_large.pth).
|
35 |
+
|
36 |
+
|
37 |
+
|
38 |
+
## Citation
|
39 |
+
|
40 |
+
```
|
41 |
+
@article{SegGPT,
|
42 |
+
title={SegGPT: Segmenting Everything In Context},
|
43 |
+
author={Wang, Xinlong and Zhang, Xiaosong and Cao, Yue and Wang, Wen and Shen, Chunhua and Huang, Tiejun},
|
44 |
+
journal={arXiv preprint arXiv:2304.03284},
|
45 |
+
year={2023}
|
46 |
+
}
|
47 |
+
```
|
48 |
+
|
49 |
+
## Contact
|
50 |
+
|
51 |
+
**We are hiring** at all levels at BAAI Vision Team, including full-time researchers, engineers and interns.
|
52 |
+
If you are interested in working with us on **foundation model, visual perception and multimodal learning**, please contact [Xinlong Wang](https://www.xloong.wang/) (`[email protected]`) and [Yue Cao](http://yue-cao.me/) (`[email protected]`).
|
53 |
+
|
seggpt_teaser.png
ADDED