# ms-marco-MiniLM-L-6-v2 ## Model description This model is a fine-tuned version of ms-marco-MiniLM-L-6-v2 for relevancy evaluation in RAG scenarios. ## Training Data The model was trained on a specialized dataset for evaluating RAG responses, containing pairs of (context, response) with relevancy labels. Dataset size: 4505 training examples, 5006 validation examples. ## Performance Metrics ``` Validation Metrics: - NDCG: 0.9996 ± 0.0001 - MAP: 0.9970 ± 0.0009 - Accuracy: 0.9766 ± 0.0033 ``` ## Usage Example ```python from sentence_transformers import CrossEncoder # Load model model = CrossEncoder('xtenzr/ms-marco-MiniLM-L-6-v2_finetuned_20241120_2220') # Prepare inputs texts = [ ["Context: {...} Query: {...}", "Response: {...}"], ] # Get predictions scores = model.predict(texts) # Returns relevancy scores [0-1] ``` ## Training procedure - Fine-tuned using sentence-transformers CrossEncoder - Trained on relevancy evaluation dataset - Optimized for RAG response evaluation