You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

BART Base Text Summarization Modeli

This model is based on the Facebook BART (Bidirectional and Auto-Regressive Transformers) architecture. BART is particularly effective when fine-tuned for text generation tasks like summarization but also works well for comprehension tasks. BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.

Model Details

Model Description

This is the model card of a ๐Ÿค— transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Architecture: [BART Base]
  • Pre-trained model: [facebook/bart-base]
  • Fine-tuned for: [Summarization]
  • License: [MIT]
  • Finetuned from model: [facebook/bart-base]

Uses

  • Installation: pip install transformers

Direct Use

Here is a simple snippet oon how to use the model directly.

Load model directly

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("ChijoTheDatascientist/summarization-model") model = AutoModelForSeq2SeqLM.from_pretrained("ChijoTheDatascientist/summarization-model")

Downloads last month
10
Safetensors
Model size
139M params
Tensor type
F32
ยท
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Space using ChijoTheDatascientist/summarization-model 1