Update README.md
Browse files
README.md
CHANGED
@@ -8,18 +8,11 @@ tags:
|
|
8 |
|
9 |
# GPT2 Zinc 87m
|
10 |
|
11 |
-
This is a GPT2 style autoregressive language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/)
|
12 |
|
13 |
The model has ~87m parameters and was trained for 175000 iterations with a batch size of 3072 to a validation loss of ~.615. This model is useful for generating druglike molecules or generating embeddings from SMILES strings
|
14 |
|
15 |
## How to use
|
16 |
-
To use, install the [transformers](https://github.com/huggingface/transformers) library:
|
17 |
-
|
18 |
-
```
|
19 |
-
pip install transformers
|
20 |
-
```
|
21 |
-
|
22 |
-
Load the model from the Huggingface Hub:
|
23 |
|
24 |
```python
|
25 |
from transformers import GPT2TokenizerFast, GPT2LMHeadModel
|
|
|
8 |
|
9 |
# GPT2 Zinc 87m
|
10 |
|
11 |
+
This is a GPT2 style autoregressive language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/).
|
12 |
|
13 |
The model has ~87m parameters and was trained for 175000 iterations with a batch size of 3072 to a validation loss of ~.615. This model is useful for generating druglike molecules or generating embeddings from SMILES strings
|
14 |
|
15 |
## How to use
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
```python
|
18 |
from transformers import GPT2TokenizerFast, GPT2LMHeadModel
|