jonathanagustin's picture
Upload README.md with huggingface_hub
bcc98c7
|
raw
history blame
9.4 kB
metadata
language: en
license: mit
model_details: |2-

          ## Abstract
          This model, 'roberta-finetuned', is a question-answering chatbot trained on the SQuAD dataset, demonstrating competency in building conversational AI using recent advances in natural language processing. It utilizes a BERT model fine-tuned for extractive question answering.

          ## Data Collection and Preprocessing
          The model was trained on the Stanford Question Answering Dataset (SQuAD), which contains over 100,000 question-answer pairs based on Wikipedia articles. The data preprocessing involved tokenizing context paragraphs and questions, truncating sequences to fit BERT's max length, and adding special tokens to mark question and paragraph segments.

          ## Model Architecture and Training
          The architecture is based on the BERT transformer model, which was pretrained on large unlabeled text corpora. For this project, the BERT base model was fine-tuned on SQuAD for extractive question answering, with additional output layers for predicting the start and end indices of the answer span.

          ## SQuAD 2.0 Dataset
          SQuAD 2.0 combines the existing SQuAD data with over 50,000 unanswerable questions written adversarially by crowdworkers to look similar to answerable ones. This version of the dataset challenges models to not only produce answers when possible but also determine when no answer is supported by the paragraph and abstain from answering.
          
intended_use: |2-

          - Answering questions from the squad_v2 dataset.
          - Developing question-answering systems within the scope of the aai520-project.
          - Research and experimentation in the NLP question-answering domain.
          
limitations_and_bias: |2-

          The model inherits limitations and biases from the 'roberta-base' model, as it was trained on the same foundational data. 
          It may underperform on questions that are ambiguous or too far outside the scope of the topics covered in the squad_v2 dataset. 
          Additionally, the model may reflect societal biases present in its training data.
          
ethical_considerations: |2-

          This model should not be used for making critical decisions without human oversight, 
          as it can generate incorrect or biased answers, especially for topics not covered in the training data. 
          Users should also consider the ethical implications of using AI in decision-making processes and the potential for perpetuating biases.
          
evaluation: |2-

          The model was evaluated on the squad_v2 dataset using various metrics. These metrics, along with their corresponding scores, 
          are detailed in the 'eval_results' section. The evaluation process ensured a comprehensive assessment of the model's performance 
          in question-answering scenarios.
          
training: |2-

          The model was trained over 4 epochs with a learning rate of 2e-05, using a batch size of 64. 
          The training utilized a cross-entropy loss function and the AdamW optimizer, with gradient accumulation over 4 steps.
          
tips_and_tricks: |2-

          For optimal performance, questions should be clear, concise, and grammatically correct. 
          The model performs best on questions related to topics covered in the squad_v2 dataset. 
          It is advisable to pre-process text for consistency in encoding and punctuation, and to manage expectations for questions on topics outside the training data.
          
model-index:
  - name: roberta-finetuned
    results:
      - task:
          type: question-answering
        dataset:
          name: SQuAD v2
          type: squad_v2
        metrics:
          - type: Exact
            value: 100
          - type: F1
            value: 100
          - type: Total
            value: 2
          - type: Hasans Exact
            value: 100
          - type: Hasans F1
            value: 100
          - type: Hasans Total
            value: 2
          - type: Best Exact
            value: 100
          - type: Best Exact Thresh
            value: 0.9603068232536316
          - type: Best F1
            value: 100
          - type: Best F1 Thresh
            value: 0.9603068232536316
          - type: Total Time In Seconds
            value: 0.034005987999989884
          - type: Samples Per Second
            value: 58.813171374423675
          - type: Latency In Seconds
            value: 0.017002993999994942

Model Card for Model ID

Model Details

Model Description

  • Developed by: [More Information Needed]
  • Shared by [optional]: [More Information Needed]
  • Model type: [More Information Needed]
  • Language(s) (NLP): en
  • License: mit
  • Finetuned from model [optional]: [More Information Needed]

Model Sources [optional]

  • Repository: [More Information Needed]
  • Paper [optional]: [More Information Needed]
  • Demo [optional]: [More Information Needed]

Uses

Direct Use

[More Information Needed]

Downstream Use [optional]

[More Information Needed]

Out-of-Scope Use

[More Information Needed]

Bias, Risks, and Limitations

[More Information Needed]

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.

How to Get Started with the Model

Use the code below to get started with the model.

[More Information Needed]

Training Details

Training Data

[More Information Needed]

Training Procedure

Preprocessing [optional]

[More Information Needed]

Training Hyperparameters

  • Training regime: [More Information Needed]

Speeds, Sizes, Times [optional]

[More Information Needed]

Evaluation

Testing Data, Factors & Metrics

Testing Data

[More Information Needed]

Factors

[More Information Needed]

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

Model Examination [optional]

[More Information Needed]

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: [More Information Needed]
  • Hours used: [More Information Needed]
  • Cloud Provider: [More Information Needed]
  • Compute Region: [More Information Needed]
  • Carbon Emitted: [More Information Needed]

Technical Specifications [optional]

Model Architecture and Objective

[More Information Needed]

Compute Infrastructure

[More Information Needed]

Hardware

[More Information Needed]

Software

[More Information Needed]

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Glossary [optional]

[More Information Needed]

More Information [optional]

[More Information Needed]

Model Card Authors [optional]

[More Information Needed]

Model Card Contact

[More Information Needed]