Kiran2004 commited on
Commit
3f389b1
1 Parent(s): c336a4a

Training in progress epoch 0

Browse files
Files changed (2) hide show
  1. README.md +17 -21
  2. tf_model.h5 +1 -1
README.md CHANGED
@@ -4,55 +4,51 @@ base_model: deepset/roberta-base-squad2
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
7
- - name: Kiran2004/my_qa_model
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
12
  probably proofread and complete it, then remove this comment. -->
13
 
14
- # Kiran2004/my_qa_model
15
-
16
- This model is a fine-tuned version of [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) on an unknown dataset.It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
17
 
 
18
  It achieves the following results on the evaluation set:
19
- - Train Loss: 3.4107
20
- - Validation Loss: 9.9990
21
- - Epoch: 1
 
 
22
 
23
- ## Usage
24
 
25
- ### In Transformers
26
- ```python
27
- from transformers import pipeline
28
 
29
- model_name = "Kiran2004/Roberta_qca_sample"
30
 
31
- question_answerer = pipeline("question-answering", model = model_name)
32
 
33
- question = "How many programming languages does BLOOM support?"
34
- context = "BLOOM has 176 billion parameters and can generate text in 46 languages natural languages and 13 programming languages."
35
 
36
- question_answerer(question=question, context=context)
37
- ```
38
 
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 8, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
43
  - training_precision: float32
44
 
45
  ### Training results
46
 
47
  | Train Loss | Validation Loss | Epoch |
48
  |:----------:|:---------------:|:-----:|
49
- | 3.0643 | 9.9990 | 0 |
50
- | 3.4107 | 9.9990 | 1 |
51
 
52
 
53
  ### Framework versions
54
 
55
  - Transformers 4.38.2
56
  - TensorFlow 2.15.0
57
- - Datasets 2.18.0
58
  - Tokenizers 0.15.2
 
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
7
+ - name: Kiran2004/Roberta_qca_sample
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
12
  probably proofread and complete it, then remove this comment. -->
13
 
14
+ # Kiran2004/Roberta_qca_sample
 
 
15
 
16
+ This model is a fine-tuned version of [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 0.4681
19
+ - Validation Loss: 0.0034
20
+ - Epoch: 0
21
+
22
+ ## Model description
23
 
24
+ More information needed
25
 
26
+ ## Intended uses & limitations
 
 
27
 
28
+ More information needed
29
 
30
+ ## Training and evaluation data
31
 
32
+ More information needed
 
33
 
34
+ ## Training procedure
 
35
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 80, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
40
  - training_precision: float32
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
+ | 0.4681 | 0.0034 | 0 |
 
47
 
48
 
49
  ### Framework versions
50
 
51
  - Transformers 4.38.2
52
  - TensorFlow 2.15.0
53
+ - Datasets 2.19.0
54
  - Tokenizers 0.15.2
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d38f634f3304793745a6b4783d20bdce84d28e9d8d919c7a3ca32225cc139b8
3
  size 496513256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8a63288800f3b7e7a86086fb91129f34b9b9d9abefccc77e5b8b7894b1e4daf
3
  size 496513256