kevinkrahn commited on
Commit
07cb18b
1 Parent(s): 7c4bf45

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -14,16 +14,22 @@ tags:
14
 
15
  ## Sentence embeddings for English and Ancient Greek
16
 
17
- The HLM model architecture is based on [Heidelberg-Boston @ SIGTYP 2024 Shared Task: Enhancing Low-Resource Language Analysis With Character-Aware Hierarchical Transformers](https://aclanthology.org/2024.sigtyp-1.16/) but uses a simpler architecture with rotary embeddings instead of using DeBERTa as a base architecture. This architecture produces superior results compared to the vanilla BERT architecture for low-resource languages like Ancient Greek. It is trained to produce sentence embeddings using the method described in [Sentence Embedding Models for Ancient Greek Using Multilingual Knowledge Distillation](https://aclanthology.org/2023.alp-1.2/).
 
 
 
 
18
 
19
  This model was distilled from `BAAI/bge-base-en-v1.5` for embedding English and Ancient Greek text.
20
 
21
  ## Usage (Sentence-Transformers)
22
 
23
- This model is currently incompatible with the latest version of the sentence-transformers library. For now, you must use this fork:
 
 
24
  https://github.com/kevinkrahn/sentence-transformers
25
 
26
- Then you can use the model like this:
27
 
28
  ```python
29
  from sentence_transformers import SentenceTransformer
 
14
 
15
  ## Sentence embeddings for English and Ancient Greek
16
 
17
+ This model creates sentence embeddings in a shared vector space for Ancient Greek and English text.
18
+
19
+ The base model uses a modified version of the HLM architecture described in [Heidelberg-Boston @ SIGTYP 2024 Shared Task: Enhancing Low-Resource Language Analysis With Character-Aware Hierarchical Transformers](https://aclanthology.org/2024.sigtyp-1.16/)
20
+
21
+ This model is trained to produce sentence embeddings using the multilingual knowledge distillation method and datasets described in [Sentence Embedding Models for Ancient Greek Using Multilingual Knowledge Distillation](https://aclanthology.org/2023.alp-1.2/).
22
 
23
  This model was distilled from `BAAI/bge-base-en-v1.5` for embedding English and Ancient Greek text.
24
 
25
  ## Usage (Sentence-Transformers)
26
 
27
+ **This model is currently incompatible with the latest version of the sentence-transformers library.**
28
+
29
+ For now, either use HuggingFace Transformers directly (see below) or the following fork of sentence-transformers:
30
  https://github.com/kevinkrahn/sentence-transformers
31
 
32
+ You can use the model with sentence-transformers like this:
33
 
34
  ```python
35
  from sentence_transformers import SentenceTransformer