MacBERTh

This model is a Historical Language Model for English coming from the MacBERTh project.

The architecture is based on BERT base uncased from the original BERT pre-training codebase. The training material comes from different sources including:

  • EEBO
  • ECCO
  • COHA
  • CLMET3.1
  • EVANS
  • Hansard Corpus

with a total word count of approximately 3.9B tokens.

Details and evaluation can be found in the accompanying publications:

Downloads last month
87
Inference API
Unable to determine this model’s pipeline type. Check the docs .