Model Card for Tokyo8 Customer Service Model (Indonesia)

This is the model card of a 🤗 transformers model based on the Llama-3.2-3B architecture, which has been finetuned by PT Clevio on a dataset of customer service conversations for the Tokyo8 product in Indonesia.

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: PT Clevio
  • Funded by : PT Clevio
  • Shared by : PT Clevio
  • Model type: Causal Language Model
  • Language(s) (NLP): English, Indonesian
  • License: Apache-2.0
  • Finetuned from model : meta-llama/Llama-3.2-3B

Model Sources [optional]

  • Repository: bagasbgs2516/llama3.2-3B-cs-Tokyo8

Uses

Direct Use

The Tokyo8 Customer Service Model (Indonesia) can be used for generating responses to customer inquiries and providing helpful information about the Tokyo8 product in the Indonesian market.

Downstream Use

This model can be integrated into customer service chatbots or virtual assistants to provide automated responses to customer questions about the Tokyo8 product in Indonesia.

Out-of-Scope Use

The model should not be used for generating content unrelated to the Tokyo8 product or for purposes outside of customer service in Indonesia. The model may also have biases or limitations that should be considered before deployment.

Bias, Risks, and Limitations

The model was trained on a dataset of customer service conversations in Indonesia, which may contain biases or reflect certain demographic or linguistic patterns. The model's performance may also be limited to the specific domain of the Tokyo8 product in Indonesia and may not generalize well to other products or domains.

Recommendations

Users should be aware of the potential biases and limitations of the model, and should carefully evaluate its performance and appropriateness for their use case. Monitoring for unintended outputs or biases is recommended, and the model should be used with appropriate safeguards and oversight.

How to Get Started with the Model

Use the code below to get started with the Tokyo8 Customer Service Model (Indonesia).

# Load model From Hugging Face Hub
model = AutoModelForCausalLM.from_pretrained("bagasbgs2516/llama3.2-3B-cs-Tokyo8")
tokenizer = AutoTokenizer.from_pretrained("bagasbgs2516/llama3.2-3B-cs-Tokyo8")

# Try for inferensi
input_text = "Halo, Apa itu Tokyo8?"
input_ids = tokenizer.encode(input_text, return_tensors="pt")

output = model.generate(input_ids, max_length=50, num_return_sequences=1)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Training Details

Training Data

The model was finetuned on a dataset of customer service conversations related to the Tokyo8 product in Indonesia, which was provided in a CSV file.

Training Procedure

Preprocessing

The dataset was preprocessed by tokenizing the text and splitting the conversations into input-response pairs.

Training Hyperparameters

  • Training regime: The model was finetuned for 6 epochs, with a batch size of 1, a gradient accumulation of 8 steps, a learning rate of 5e-4, and a weight decay of 0.01.

Speeds, Sizes, Times

The finetuning process took approximately 1.5 hours on a NVIDIA T4 GPU with 15GB of memory.

Evaluation

Testing Data, Factors & Metrics

Testing Data

The model was evaluated on a held-out portion of the customer service conversation dataset, which was not used for training.

Factors

The evaluation considered factors such as the quality and relevance of the generated responses, as well as the model's ability to handle different types of customer inquiries in the Indonesian market.

Metrics

The model's performance was evaluated using metrics such as perplexity, BLEU score, and human evaluation. The model achieved a BLEU score of 0.4227 and the precision scores for different n-gram levels ranged from 0.1419 to 0.2038.

Results

The Tokyo8 Customer Service Model (Indonesia) achieved a BLEU score of 0.4227 and demonstrated promising performance in generating relevant and helpful responses to customer inquiries based on the evaluation metrics.

Summary

The Tokyo8 Customer Service Model (Indonesia) has been successfully finetuned by PT Clevio on a dataset of customer service conversations related to the Tokyo8 product in the Indonesian market. The model demonstrates good performance in generating relevant and helpful responses to customer inquiries, but may have biases or limitations that should be carefully considered before deployment.

Model Architecture and Objective

The Tokyo8 Customer Service Model (Indonesia) is based on the Llama-3.2-3B architecture, which is a large language model trained for causal language modeling. The objective of the finetuning process was to adapt the model to the specific domain of customer service conversations for the Tokyo8 product in the Indonesian market.

Compute Infrastructure

The model was finetuned on a system with 1 NVIDIA T4 GPU with 15GB of memory, 51GB of RAM.

Hardware

  • NVIDIA T4 GPU (15GB)
  • RAM 51GB

Software

  • PyTorch
  • Transformers library
  • Peft

Citation

BibTeX:

@misc{tokyo8-customer-service-model-indonesia,
    title={Tokyo8 Customer Service Model (Indonesia)},
    author={PT Clevio},
    year={2024},
    howpublished={\url{https://huggingface.co/bagasbgs2516/llama3.2-3B-cs-Tokyo8}},
}

APA:

PT Clevio. (2024). Tokyo8 Customer Service Model (Indonesia). Hugging Face. https://huggingface.co/bagasbgs2516/llama3.2-3B-cs-Tokyo8

Model Card Contact

For more information or inquiries about the Tokyo8 Customer Service Model (Indonesia), please contact: Bagas Tri Adiwira [email protected] PT Clevio

Downloads last month
27
GGUF
Model size
3.21B params
Architecture
llama
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for bagasbgs2516/llama3.2-3B-cs-Tokyo8

Quantized
(45)
this model