tianshuzhang
commited on
Commit
·
342f365
1
Parent(s):
f3109a2
Update README.md
Browse files
README.md
CHANGED
@@ -1,13 +1,11 @@
|
|
1 |
---
|
2 |
license: cc-by-4.0
|
3 |
-
---
|
4 |
-
|
5 |
-
datasets:
|
6 |
-
- osunlp/TableInstruct
|
7 |
language:
|
8 |
- en
|
9 |
---
|
10 |
-
|
|
|
|
|
11 |
|
12 |
Project Page: [https://osu-nlp-group.github.io/TableLlama/](https://osu-nlp-group.github.io/TableLlama/)
|
13 |
|
@@ -20,7 +18,9 @@ Code: [https://osu-nlp-group.github.io/TableLlama/](https://osu-nlp-group.github
|
|
20 |
We introduce TableLlama, an open-source large generalist model specifically tailored for various table-based tasks. The TableLlama model is trained on 🤗 [TableInstruct Dataset](https://huggingface.co/datasets/osunlp/TableInstruct), a meticulously curated instruction tuning dataset for tables. TableLlama is tuned on 2.6 million table-based task data, and can handle up to 8K context!
|
21 |
|
22 |
|
23 |
-
## Model
|
|
|
|
|
24 |
|
25 |
|
26 |
## Training Data
|
@@ -55,10 +55,6 @@ appropriately completes the request.
|
|
55 |
### Response:
|
56 |
```
|
57 |
|
58 |
-
```
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
## Limitations
|
63 |
We've tried our best to build table generalist models. However, we acknowledge that the models' performance may vary based on the complexity and specifics of the table tasks and datasets. Still not all tablep-based tasks can be covered comprehensively.
|
64 |
|
@@ -75,4 +71,4 @@ If you use the models, data, or code from this project, please cite the original
|
|
75 |
archivePrefix={arXiv},
|
76 |
primaryClass={cs.CL}
|
77 |
}
|
78 |
-
```
|
|
|
1 |
---
|
2 |
license: cc-by-4.0
|
|
|
|
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
---
|
6 |
+
|
7 |
+
---
|
8 |
+
# TableLlama: Towards Open Large Generalist Models for Tables
|
9 |
|
10 |
Project Page: [https://osu-nlp-group.github.io/TableLlama/](https://osu-nlp-group.github.io/TableLlama/)
|
11 |
|
|
|
18 |
We introduce TableLlama, an open-source large generalist model specifically tailored for various table-based tasks. The TableLlama model is trained on 🤗 [TableInstruct Dataset](https://huggingface.co/datasets/osunlp/TableInstruct), a meticulously curated instruction tuning dataset for tables. TableLlama is tuned on 2.6 million table-based task data, and can handle up to 8K context!
|
19 |
|
20 |
|
21 |
+
## Model
|
22 |
+
|
23 |
+
[TableLlama-7B](https://huggingface.co/osunlp/TableLlama/)
|
24 |
|
25 |
|
26 |
## Training Data
|
|
|
55 |
### Response:
|
56 |
```
|
57 |
|
|
|
|
|
|
|
|
|
58 |
## Limitations
|
59 |
We've tried our best to build table generalist models. However, we acknowledge that the models' performance may vary based on the complexity and specifics of the table tasks and datasets. Still not all tablep-based tasks can be covered comprehensively.
|
60 |
|
|
|
71 |
archivePrefix={arXiv},
|
72 |
primaryClass={cs.CL}
|
73 |
}
|
74 |
+
```
|