This is a experimental llama2 7B qlora made using the VNTL-v2-2k-small dataset. Unlike version 0.1, the input was masked out of the loss calculation during training.

The objectives of this fine-tune are:

  1. Teaching the model how to translate while respecting the context.
  2. Teaching the model how to translate while following the character's metadata.
  3. Teaching the model how to translate while respecting the translation fidelity.

I can say with certainty that objectives 1 and 2 were completed successfully. However, the objective 3 wasn't, most likely due to the difficulty of the model making such association, so instead, I used the fidelity classification just to exclude the translations with low/medium fidelity from affecting the loss calculation, since these translations are, most of the time, creative translations or straight up mismatched translations.

This is an prompt example:

<<START>>
Name: Uryuu Shingo (η“œη”Ÿ 新吾) | Gender: Male | Aliases: Onii-chan (γŠε…„γ‘γ‚ƒγ‚“)
Name: Uryuu Sakuno (η“œη”Ÿ ζ‘œδΉƒ) | Gender: Female | Aliases: None
<<JAPANESE>>
γ€ζ‘œδΉƒγ€‘οΌšγ€Žβ€¦β€¦γ”γ‚γ‚“γ€
<<ENGLISH>> (fidelity = absolute)
【Sakunoγ€‘οΌšγ€Ž... Sorry.』
<<JAPANESE>>
γ€ζ–°εΎγ€‘οΌšγ€Œγ†γ†γ‚“γ€γ“γ†θ¨€γ£γ‘γ‚ƒγͺγ‚“γ γ‘γ©γ€θΏ·ε­γ§γ‚ˆγ‹γ£γŸγ‚ˆγ€‚ζ‘œδΉƒγ―ε―ζ„›γ„γ‹γ‚‰γ€γ„γ‚γ„γ‚εΏƒι…γ—γ‘γ‚ƒγ£γ¦γŸγ‚“γ γžδΏΊγ€
<<ENGLISH>> (fidelity = high)

The generated translation for that prompt, with temperature 0, is:

【Shingoγ€‘οΌšγ€ŒDon't worry about it. I was just glad you were lost. You're cute, so I was worried about you.」
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Examples
Unable to determine this model's library. Check the docs .

Dataset used to train lmg-anon/vntl-7b-v0.2-qlora