fbeawels commited on
Commit
6eab84f
1 Parent(s): cdd6331

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -31
README.md CHANGED
@@ -1,60 +1,83 @@
 
1
  license: mit
2
- base_model: microsoft/Phi-3-medium-128k-instruct
3
  library_name: adapters
4
  datasets:
5
- - awels/ocpvirt_admin_dataset
6
  language:
7
- - en
8
  widget:
9
- - text: Who are you, Thready ?
10
  tags:
11
- - awels
12
- - redhat
13
- Thready Model Card
14
- Model Details
15
- Model Name: Thready
16
 
17
- Model Type: Transformer-based leveraging Microsoft Phi 14b 128k tokens
18
 
19
- Publisher: Awels Engineering
 
20
 
21
- License: MIT
22
 
23
- Model Description: Thready is a sophisticated model designed to help as an AI agent focusing on the Red Hat Openshift Virtualization solution. It leverages advanced machine learning techniques to provide efficient and accurate solutions. It has been trained on the full docments corpus of OCP Virt 4.16.
24
 
25
- Dataset
26
- Dataset Name: awels/ocpvirt_admin_dataset
27
 
28
- Dataset Source: Hugging Face Datasets
 
29
 
30
- Dataset License: MIT
 
31
 
32
- Dataset Description: The dataset used to train Thready consists of all the public documents available on Red Hat Openshift Virtualization. This dataset is curated to ensure a comprehensive representation of typical administrative scenarios encountered in Openshift Virtualization.
33
 
34
- Training Details
35
- Training Data: The training data includes 70,000 Questions and Answers generated by the Bonito LLM. The dataset is split into 3 sets of data (training, test and validation) to ensure robust model performance.
36
 
37
- Training Procedure: Thready was trained using supervised learning with cross-entropy loss and the Adam optimizer. The training involved 1 epoch, a batch size of 4, a learning rate of 5.0e-06, and a cosine learning rate scheduler with gradient checkpointing for memory efficiency.
 
38
 
39
- Hardware: The model was trained on a single NVIDIA H100 SXM graphic card.
40
 
41
- Framework: The training was conducted using PyTorch.
 
42
 
43
- Evaluation
44
- Evaluation Metrics: Thready was evaluated on the training dataset:
45
 
46
- epoch = 1.0 total_flos = 74851620GF train_loss = 2.6706 train_runtime = 0:41:52.37 train_samples_per_second = 22.229 train_steps_per_second = 5.55epoch = 1.0
 
 
 
 
 
 
 
 
 
 
 
47
  total_flos = 273116814GF
48
  train_loss = 1.5825
49
  train_runtime = 1:33:44.28
50
  train_samples_per_second = 9.803
51
  train_steps_per_second = 2.451
52
 
53
- Performance: The model achieved the following results on the evaluation dataset:
 
 
 
 
 
 
 
 
 
54
 
55
- epoch = 1.0 eval_loss = 2.2243 eval_runtime = 0:02:21.35 eval_samples = 11191 eval_samples_per_second = 97.867 eval_steps_per_second = 24.47
56
 
57
- Intended Use
58
- Primary Use Case: Thready is intended to be used locally in an agent swarm to colleborate together to solve Red Hat Openshift Virtualization related problems.
59
 
60
- Limitations: While Thready is highly effective, it may have limitations due to the model size. An 8b model based on Llama 3 is used internally at Awels Engineering.
 
 
1
+ ---
2
  license: mit
3
+ base_model: microsoft/Phi-3-mini-128k-instruct
4
  library_name: adapters
5
  datasets:
6
+ - awels/ocpvirt_admin_dataset
7
  language:
8
+ - en
9
  widget:
10
+ - text: Who are you, Thready ?
11
  tags:
12
+ - awels
13
+ - redhat
14
+ ---
 
 
15
 
16
+ # Thready Model Card
17
 
18
+ ## Model Details
19
+ **Model Name:** Thready
20
 
21
+ **Model Type:** Transformer-based leveraging Microsoft Phi 3b 128k tokens
22
 
23
+ **Publisher:** Awels Engineering
24
 
25
+ **License:** MIT
 
26
 
27
+ **Model Description:**
28
+ Thready is a sophisticated model designed to help as an AI agent focusing on the Red Hat Openshift Virtualization solution. It leverages advanced machine learning techniques to provide efficient and accurate solutions. It has been trained on the full docments corpus of OCP Virt 4.16.
29
 
30
+ ## Dataset
31
+ **Dataset Name:** [awels/ocpvirt_admin_dataset](https://huggingface.co/datasets/awels/ocpvirt_admin_dataset)
32
 
33
+ **Dataset Source:** Hugging Face Datasets
34
 
35
+ **Dataset License:** MIT
 
36
 
37
+ **Dataset Description:**
38
+ The dataset used to train Thready consists of all the public documents available on Red Hat Openshift Virtualization. This dataset is curated to ensure a comprehensive representation of typical administrative scenarios encountered in Openshift Virtualization.
39
 
40
+ ## Training Details
41
 
42
+ **Training Data:**
43
+ The training data includes 70,000 Questions and Answers generated by the [Bonito LLM](https://github.com/BatsResearch/bonito). The dataset is split into 3 sets of data (training, test and validation) to ensure robust model performance.
44
 
45
+ **Training Procedure:**
46
+ Thready was trained using supervised learning with cross-entropy loss and the Adam optimizer. The training involved 1 epoch, a batch size of 4, a learning rate of 5.0e-06, and a cosine learning rate scheduler with gradient checkpointing for memory efficiency.
47
 
48
+ **Hardware:**
49
+ The model was trained on a single NVIDIA H100 SXM graphic card.
50
+
51
+ **Framework:**
52
+ The training was conducted using PyTorch.
53
+
54
+ ## Evaluation
55
+
56
+ **Evaluation Metrics:**
57
+ Thready was evaluated on the training dataset:
58
+
59
+ > epoch = 1.0
60
  total_flos = 273116814GF
61
  train_loss = 1.5825
62
  train_runtime = 1:33:44.28
63
  train_samples_per_second = 9.803
64
  train_steps_per_second = 2.451
65
 
66
+ **Performance:**
67
+ The model achieved the following results on the evaluation dataset:
68
+
69
+ > epoch = 1.0
70
+ eval_loss = 1.3341
71
+ eval_runtime = 0:04:02.02
72
+ eval_samples = 11191
73
+ eval_samples_per_second = 56.469
74
+ eval_steps_per_second = 14.118
75
+
76
 
77
+ ## Intended Use
78
 
79
+ **Primary Use Case:**
80
+ Thready is intended to be used locally in an agent swarm to colleborate together to solve Red Hat Openshift Virtualization related problems.
81
 
82
+ **Limitations:**
83
+ This 14b model is an upscale of the 3b model. Much better loss than the 3b so results should be better.