--- license: mit language: - en metrics: - accuracy widget: - text: "You have the right to use CommunityConnect for its intended purpose of connecting with others, sharing content responsibly, and engaging in constructive dialogue. You are responsible for the content you post and must respect the rights and privacy of others." example_title: "Fair Clause" - text: " We reserve the right to suspend, terminate, or restrict your access to the platform at any time and for any reason, without prior notice or explanation. This includes but is not limited to violations of our community guidelines or terms of service, as determined solely by ConnectWorld." example_title: "Unfair Clause" library_name: transformers pipeline_tag: text-classification tags: - nlp - bert - TOS --- # BertTOS v2: Terms of Service Unfairness Classifier ## Model Details - **Model Name:** BertTOS v2 - **Model Type:** Fine-tuned BERT for sequence classification - **Version:** 2.0 - **Language(s):** English - **License:** [MIT] - **Developer:** [Himanshu Mohanty] ## Model Description BertTOS v2 is a fine-tuned BERT model designed to classify clauses in Terms of Service (ToS) documents based on their unfairness level. This model can help users identify potentially problematic clauses in legal documents, particularly in the context of consumer protection. ### Task The model performs multi-class classification on individual sentences or clauses, categorizing them into three levels of unfairness: 0. Clearly Fair 1. Potentially Unfair 2. Clearly Unfair ### Training Data The model was trained on the [CodeHima/TOS_Dataset](https://huggingface.co/datasets/CodeHima/TOS_Dataset) dataset, which contains annotated sentences from Terms of Service documents. Each sentence is labeled with one of the three unfairness levels. ### Model Architecture - Base Model: BERT (bert-base-uncased) - Fine-tuning: Sequence classification head - Input: Tokenized text (max length 512 tokens) - Output: Probabilities for each unfairness level ## Performance The model's performance metrics on the test set: - Accuracy: [0.8795761078998073] - F1 Score (weighted): [0.885282] - Precision (weighted): [0.883729] - Recall (weighted): [0.889157] ## Limitations - The model is trained on English language ToS documents and may not perform well on other languages or legal contexts. - Performance may vary depending on the specific wording and context of clauses. - The model should be used as a tool to assist human judgment, not as a definitive legal assessment. ## Ethical Considerations - This model is intended to help identify potentially unfair clauses, but it should not be considered as legal advice. - Users should be aware of potential biases in the training data and model predictions. - The model's output should be reviewed by legal professionals for critical applications. ## How to Use You can use this model directly with the Hugging Face `transformers` library: ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch # Load model and tokenizer model_name = "YourHuggingFaceUsername/TOSBertV2" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) # Function to predict unfairness level def predict_unfairness(text): inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512) model.eval() with torch.no_grad(): outputs = model(**inputs) probabilities = torch.softmax(outputs.logits, dim=-1).squeeze() predicted_class = torch.argmax(probabilities).item() label_mapping = {0: 'clearly_fair', 1: 'potentially_unfair', 2: 'clearly_unfair'} predicted_label = label_mapping[predicted_class] return predicted_label, probabilities.tolist() # Example usage clause = "The company reserves the right to change these terms at any time without notice." predicted_label, probabilities = predict_unfairness(clause) print(f"Predicted unfairness level: {predicted_label}") print("Probabilities:") for label, prob in zip(['clearly_fair', 'potentially_unfair', 'clearly_unfair'], probabilities): print(f"{label}: {prob:.4f}") ``` ## Training The model was trained using the following hyperparameters: - Epochs: 3 - Batch Size: 16 - Learning Rate: [ ] - Optimizer: AdamW - Weight Decay: 0.01 ## Citation If you use this model in your research, please cite: ```bibtex @misc{TOSBertV2, author = {Himanshu Mohanty}, title = {TOSBertV2: is a fine-tuned BERT model designed to classify clauses in Terms of Service}, year = {2024}, publisher = {Hugging Face}, journal = {Hugging Face Model Hub}, howpublished = {\url{https://huggingface.co/CodeHima/TOSBertV2}} }