--- language: - en tags: - deval - evaluation - llama library_name: transformers model-index: - name: roadz/dv-finetuned-211124 results: [] pipeline_tag: text-generation --- # Model Card for roadz/dv-finetuned-211124 This model is fine-tuned for evaluating LLM outputs in RAG scenarios, focusing on: - Hallucination detection - Attribution accuracy - Summary completeness - Response relevancy ## Model Details ### Model Architecture - Base Model: LLaMA-3.1-8B - Architecture Type: llama - Parameters: Not specified - Training Type: Fine-tuned for evaluation ### Hardware Requirements - Minimum GPU Memory: 16GB - Recommended GPU Memory: 24GB - Format: SafeTensors ## Usage This model is designed for the De-Val subnet and requires specific pipeline code for evaluation tasks. ### Generation Configuration - Max Length: Not specified - Temperature: 0.6 - Top-p: 0.9 - Top-k: 50 ## Training The model was fine-tuned on evaluation tasks including: - Hallucination detection scenarios - Attribution verification tasks - Summary completeness assessment - Response relevancy evaluation ## Limitations - Designed specifically for evaluation tasks - Requires De-Val pipeline code - Not intended for general text generation ## Last Updated 2024-11-21