File size: 3,307 Bytes
a560a5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 |
# LLaVA-Plus Server

As shown in the figure above, we need to build an API cloud to main all tools.
## Launch a controller
Build a controller like the guides in [readme](README.md#1-Launch-a-controller).
Our tool workers and llm-model workers share the same controller.
## Launch tool workers: An Example
We provide all model workers in the `serve/` folder. The tool worker is named as `{tool_name}_worker.py`. To run a worker, make sure you have installed the required tool(or copy the worker file to different tool folders).
An example for the Grounded-SAM, whch includes Grounding-DINO, SAM, and Grounded-SAM.
```sh
git clone https://github.com/IDEA-Research/Grounded-Segment-Anything
python -m pip install -e GroundingDINO
python -m pip install -e segment_anything
python serve/grounding_dino_worker.py
```
Wait until the process finishes loading the model and you see "Uvicorn running on ...". You need to open another terminal process for other operations.
Test the worker:
```sh
python serve/grounding_dino_test_message.py
```
## All Tools
| Tool Name | Install Source |
| --------------- | -------------------------------------------------------------------------------------------------------------------------------------------------- |
| Grounding DINO | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) |
| SAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) |
| Grounded-SAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) |
| CLIP-Retrieval | [https://github.com/rom1504/clip-retrieval](https://github.com/rom1504/clip-retrieval) |
| InstructPix2Pix | [https://github.com/huggingface/diffusers](https://github.com/huggingface/diffusers) |
| StableDiffusion | [https://github.com/huggingface/diffusers](https://github.com/huggingface/diffusers) |
| BLIP2 | [https://github.com/huggingface/transformers](https://github.com/huggingface/transformers) |
| RAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) |
| Semantic-SAM | [https://github.com/UX-Decoder/Semantic-SAM](https://github.com/UX-Decoder/Semantic-SAM) |
| SEEM | [https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once](https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once) |
| OCR(easyocr) | [https://github.com/JaidedAI/EasyOCR](https://github.com/JaidedAI/EasyOCR) | |