|
--- |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
--- |
|
# GraphGPT |
|
GraphGPT is a graph-oriented Large Language Model tuned by Graph Instruction Tuning paradigm. |
|
|
|
## Model Details |
|
GraphGPT is a graph-oriented Large Language Model tuned by Graph Instruction Tuning paradigm based on the [Vicuna-7B-v1.5 model](https://huggingface.co/lmsys/vicuna-7b-v1.5). |
|
* Developed by: [Data Intelligence Lab](https://sites.google.com/view/chaoh/group-join-us)@HKU |
|
* Model type: An auto-regressive language model based on the transformer architecture. |
|
* Finetuned from model: [Vicuna-7B-v1.5 model](https://huggingface.co/lmsys/vicuna-7b-v1.5). |
|
## Model Sources |
|
* Repository: [https://github.com/HKUDS/GraphGPT](https://github.com/HKUDS/GraphGPT) |
|
* Paper: []() |
|
* Project: [https://graphgpt.github.io/](https://graphgpt.github.io/) |
|
## Uses |
|
This version of GraphGPT is tuned utilizing the mixing instruction data, which is able to handle both node classification and link prediction for different graph datasets. |
|
## How to Get Started with the Model |
|
* Command line interface: Plaese refer to [https://github.com/HKUDS/GraphGPT](https://github.com/HKUDS/GraphGPT) to evaluate our GraphGPT. |
|
* Gradio demo is under development. |