summarization-model / README.md
ChijoTheDatascientist's picture
Update README.md
2f1750d verified
metadata
library_name: transformers
license: mit
language:
  - en
pipeline_tag: summarization

BART Base Text Summarization Modeli

This model is based on the Facebook BART (Bidirectional and Auto-Regressive Transformers) architecture. BART is particularly effective when fine-tuned for text generation tasks like summarization but also works well for comprehension tasks. BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.

Model Details

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Architecture: [BART Base]
  • Pre-trained model: [facebook/bart-base]
  • Fine-tuned for: [Summarization]
  • License: [MIT]
  • Finetuned from model: [facebook/bart-base]

Uses

  • Installation: pip install transformers

Direct Use

Here is a simple snippet oon how to use the model directly.

Load model directly

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("ChijoTheDatascientist/summarization-model") model = AutoModelForSeq2SeqLM.from_pretrained("ChijoTheDatascientist/summarization-model")