shafire commited on
Commit
e4967c8
1 Parent(s): 585f081

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -6
README.md CHANGED
@@ -22,12 +22,58 @@ This model is ideal for advanced NLP tasks, including ethical decision-making, m
22
 
23
  **Model Overview**
24
 
25
- - **Developer**: Shafaet Brady Hussain - [ResearchForum](https://researchforum.online)
26
- - **Funded by**: [Researchforum.online](https://researchforum.online)
27
- - **Language**: English
28
- - **Model Type**: Causal Language Model
29
- - **Base Model**: LLaMA 3.1 8B (Meta)
30
- - **License**: Apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
  **Usage**: Run on any web interface or as a bot for self-hosted solutions. Designed to run smoothly on CPU.
33
 
 
22
 
23
  **Model Overview**
24
 
25
+ SpectraMind Models
26
+ A collection of fine-tuned Llama models optimized for CPU performance using the GGUF format. These models are designed for efficient inference with llama.cpp and other lightweight environments.
27
+
28
+ Model Directory
29
+ 1. MicroSpectraMind (1B Model)
30
+ Base Model: Fine-tuned from Llama-3.2-1B.
31
+ Fine-Tuning Details: Explain the dataset used (e.g., domain-specific text, chat dialogues, or tasks such as summarization or question answering).
32
+ Optimization: Quantized to both:
33
+ f16: For maximum accuracy.
34
+ q8_0: For reduced size and faster inference.
35
+ Use Case: Ideal for lightweight applications such as embedded systems or single-threaded inference on CPUs.
36
+ File Sizes:
37
+ MicroSpectraMind_f16.gguf: 2.4 GB
38
+ MicroSpectraMind_q8.gguf: 1.3 GB
39
+ 2. SpectraMind3 (3B Model)
40
+ Base Model: Fine-tuned from Llama-3.2-3B.
41
+ Fine-Tuning Details: Include key aspects of the fine-tuning, such as datasets or hyperparameters used, and what tasks it excels at.
42
+ Optimization:
43
+ f16: For higher accuracy.
44
+ q8_0: For better efficiency.
45
+ Use Case: Balances accuracy and performance, suited for general-purpose natural language tasks.
46
+ File Sizes:
47
+ SpectraMind3_f16.gguf: 4.7 GB
48
+ SpectraMind3_q8.gguf: 3.4 GB
49
+ 3. SpectraMindZ (8B Model)
50
+ Base Model: Fine-tuned from Llama-3.2-8B.
51
+ Fine-Tuning Details: Provide specifics on dataset/task for fine-tuning.
52
+ Optimization:
53
+ f16: For maximum model precision.
54
+ q8_0: For efficient deployment with minimal performance impact.
55
+ Use Case: Best for complex tasks requiring higher reasoning or multitasking.
56
+ Expected File Sizes:
57
+ SpectraMindZ_f16.gguf: Approximately 12 GB
58
+ SpectraMindZ_q8.gguf: Approximately 8 GB
59
+ Optimization and Compatibility
60
+ All models are converted to GGUF format using llama.cpp, making them optimized for CPU-based inference. These models are ideal for systems with limited resources, such as desktops, laptops, and embedded devices.
61
+ Quantized versions (q8_0) are significantly smaller and faster, while maintaining reasonable accuracy.
62
+ How to Use
63
+ Download the GGUF Files: Use the provided links to download the .gguf files.
64
+ Run on llama.cpp: Example command for inference:
65
+ bash
66
+ Copy code
67
+ ./main -m SpectraMind3_q8.gguf -p "Your prompt here"
68
+ Choose Quantization Based on Use Case:
69
+ Use f16 for maximum accuracy (e.g., research or high-precision tasks).
70
+ Use q8_0 for faster inference (e.g., real-time applications).
71
+ Model Comparison
72
+ Model Parameters f16 Size q8_0 Size Use Case
73
+ MicroSpectraMind 1B 2.4 GB 1.3 GB Lightweight, quick responses
74
+ SpectraMind3 3B 4.7 GB 3.4 GB Balanced accuracy/performance
75
+ SpectraMindZ 8B 12 GB 8 GB Advanced tasks, complex reasoning
76
+
77
 
78
  **Usage**: Run on any web interface or as a bot for self-hosted solutions. Designed to run smoothly on CPU.
79