
National Library of Sweden / KBLab
AI & ML interests
NLP, language models, acoustic models, NER
Recent Activity
KBLab is a national research infrastructure for digital humanities and social science at the National Library of Sweden. We train large language models and speech recognition models on Swedish data from the librarys collections. Check out our blog!
Some of our most popular models are
- the original KB-BERT
- a Sentence-BERT
- a Whisper model, KB-Whisper
- an NER model trained on a mix of cased and uncased data
- a wav2vec model, VoxRex
- a BART model
- a large BERT trained with Megatron-LM
Additionally we provide previous checkpoints for some models and other models closely related to those above. Models related to each other are linked in the respective model cards.
If you are unsure which model fits you best feel free to contact us.
We also provide a variant of the venerable SUC 3.0 NER-dataset extended by Språkbanken-Text dubbed SUCX 3.0 - NER and our own Swedish/Norwegian/Danish GLUE-imitation dubbed ÖverLim.
models
75

KBLab/kb-whisper-tiny

KBLab/kb-whisper-base

KBLab/kb-whisper-small

KBLab/kb-whisper-medium

KBLab/kb-whisper-large

KBLab/wav2vec2-large-voxrex-swedish

KBLab/wav2vec2-large-voxpopuli-sv-swedish

KBLab/sentence-bert-swedish-cased

KBLab/swedish-ocr-correction
