mrstarkng's picture
Update README.md
a5b3c61 verified
---
license: bsd-3-clause
---
# Hong Lou Meng Fine-tuned Model for Word Alignment
This repository contains a fine-tuned version of the **BERT multilingual model** (`bert-base-multilingual-cased`) on the **Hong Lou Meng** dataset for word alignment tasks. This model is fine-tuned using the [awesome-align](https://github.com/neulab/awesome-align) framework and is designed for Chinese-Vietnamese (Zh-Vn) alignment.
## Model Details
- **Base Model:** `bert-base-multilingual-cased`
- **Fine-tuned Dataset:** Excerpts from the classic "Hong Lou Meng" novel, annotated with Chinese and Vietnamese sentence pairs.
- **Alignment Task:** Fine-tuned to align word pairs in parallel texts for translation and linguistic analysis.
---
## Example Usage
Below is an example of how to use this model for word alignment using the `transformers` library:
```python
from transformers import AutoTokenizer, AutoModel
import torch
# Load model and tokenizer
model_name = "username/zh-vn-hongloumeng-align"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModel.from_pretrained(model_name)
# Input sentences (Chinese and Vietnamese)
source_sentence = "第一回 甄士隱夢幻識通靈 賈雨村風塵懷閨秀"
target_sentence = "Hồi thứ nhất: Chân Sĩ Ẩn mộng ảo ngộ đá thiêng, Giả Vũ Thôn phong trần nhớ giai nhân."
# Tokenize inputs
inputs = tokenizer(source_sentence, target_sentence, return_tensors="pt", padding=True, truncation=True)
# Pass through model
outputs = model(**inputs)
# Further processing for alignment visualization or analysis would follow
print("Model outputs:", outputs)