You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

BART Base Text Summarization Modeli

This model is based on the Facebook BART (Bidirectional and Auto-Regressive Transformers) architecture. BART is particularly effective when fine-tuned for text generation tasks like summarization but also works well for comprehension tasks. BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.

Model Details

Model Description

This is the model card of a ๐Ÿค— transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Architecture: [BART Base]
  • Pre-trained model: [facebook/bart-base]
  • Fine-tuned for: [Summarization]
  • License: [MIT]
  • Finetuned from model: [facebook/bart-base]

Uses

  • Installation: pip install transformers

Direct Use

Here is a simple snippet oon how to use the model directly.

Load model directly

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("ChijoTheDatascientist/summarization-model") model = AutoModelForSeq2SeqLM.from_pretrained("ChijoTheDatascientist/summarization-model")

Downloads last month
108
Safetensors
Model size
139M params
Tensor type
F32
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Space using ChijoTheDatascientist/summarization-model 1