|
# LLaVA-Plus Server |
|
|
|
 |
|
|
|
As shown in the figure above, we need to build an API cloud to main all tools. |
|
|
|
## Launch a controller |
|
Build a controller like the guides in [readme](README.md#1-Launch-a-controller). |
|
|
|
Our tool workers and llm-model workers share the same controller. |
|
|
|
## Launch tool workers: An Example |
|
We provide all model workers in the `serve/` folder. The tool worker is named as `{tool_name}_worker.py`. To run a worker, make sure you have installed the required tool(or copy the worker file to different tool folders). |
|
|
|
An example for the Grounded-SAM, whch includes Grounding-DINO, SAM, and Grounded-SAM. |
|
|
|
```sh |
|
git clone https://github.com/IDEA-Research/Grounded-Segment-Anything |
|
python -m pip install -e GroundingDINO |
|
python -m pip install -e segment_anything |
|
python serve/grounding_dino_worker.py |
|
``` |
|
|
|
Wait until the process finishes loading the model and you see "Uvicorn running on ...". You need to open another terminal process for other operations. |
|
|
|
Test the worker: |
|
```sh |
|
python serve/grounding_dino_test_message.py |
|
``` |
|
|
|
## All Tools |
|
|
|
| Tool Name | Install Source | |
|
| --------------- | -------------------------------------------------------------------------------------------------------------------------------------------------- | |
|
| Grounding DINO | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) | |
|
| SAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) | |
|
| Grounded-SAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) | |
|
| CLIP-Retrieval | [https://github.com/rom1504/clip-retrieval](https://github.com/rom1504/clip-retrieval) | |
|
| InstructPix2Pix | [https://github.com/huggingface/diffusers](https://github.com/huggingface/diffusers) | |
|
| StableDiffusion | [https://github.com/huggingface/diffusers](https://github.com/huggingface/diffusers) | |
|
| BLIP2 | [https://github.com/huggingface/transformers](https://github.com/huggingface/transformers) | |
|
| RAM | [https://github.com/IDEA-Research/Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) | |
|
| Semantic-SAM | [https://github.com/UX-Decoder/Semantic-SAM](https://github.com/UX-Decoder/Semantic-SAM) | |
|
| SEEM | [https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once](https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once) | |
|
| OCR(easyocr) | [https://github.com/JaidedAI/EasyOCR](https://github.com/JaidedAI/EasyOCR) | |