jhsmith commited on
Commit
6423609
·
1 Parent(s): 1e2d5fa

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -46,7 +46,7 @@ The model was trained with the parameters:
46
 
47
  **DataLoader**:
48
 
49
- `torch.utils.data.dataloader.DataLoader` of length 2688 with parameters:
50
  ```
51
  {'batch_size': 10, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
52
  ```
@@ -61,8 +61,8 @@ The model was trained with the parameters:
61
  Parameters of the fit()-Method:
62
  ```
63
  {
64
- "epochs": 3,
65
- "evaluation_steps": 500,
66
  "evaluator": "sentence_transformers.evaluation.TripletEvaluator.TripletEvaluator",
67
  "max_grad_norm": 1,
68
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
@@ -71,7 +71,7 @@ Parameters of the fit()-Method:
71
  },
72
  "scheduler": "warmuplinear",
73
  "steps_per_epoch": null,
74
- "warmup_steps": 806,
75
  "weight_decay": 0.01
76
  }
77
  ```
 
46
 
47
  **DataLoader**:
48
 
49
+ `torch.utils.data.dataloader.DataLoader` of length 1344 with parameters:
50
  ```
51
  {'batch_size': 10, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
52
  ```
 
61
  Parameters of the fit()-Method:
62
  ```
63
  {
64
+ "epochs": 1,
65
+ "evaluation_steps": 200,
66
  "evaluator": "sentence_transformers.evaluation.TripletEvaluator.TripletEvaluator",
67
  "max_grad_norm": 1,
68
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
 
71
  },
72
  "scheduler": "warmuplinear",
73
  "steps_per_epoch": null,
74
+ "warmup_steps": 134,
75
  "weight_decay": 0.01
76
  }
77
  ```
eval/triplet_evaluation_results_results.csv CHANGED
@@ -1,19 +1,11 @@
1
  epoch,steps,accuracy_cosinus,accuracy_manhattan,accuracy_euclidean
2
- 0,500,0.7489212513484358,0.7471233369291622,0.7489212513484358
3
- 0,1000,0.5784789644012945,0.3888888888888889,0.5783890686803308
4
- 0,1500,0.45433297375044945,0.4502876663070838,0.45532182668105
5
- 0,2000,0.4867853290183387,0.48471772743617403,0.4873247033441208
6
- 0,2500,0.47312117943185905,0.4617044228694714,0.4729413879899317
7
- 0,-1,0.47105357784969437,0.4713232650125854,0.4717727436174038
8
- 1,500,0.4842682488313556,0.48903272204243076,0.48462783171521034
9
- 1,1000,0.4874145990650845,0.4828299172959367,0.4869651204602661
10
- 1,1500,0.48759439050701187,0.49154980222941386,0.48759439050701187
11
- 1,2000,0.48768428622797555,0.48570658036677455,0.48723480762315713
12
- 1,2500,0.48894282632146707,0.490381157856886,0.4892125134843581
13
- 1,-1,0.4892125134843581,0.49065084501977707,0.4894822006472492
14
- 2,500,0.4894822006472492,0.49460625674217906,0.48975188781014023
15
- 2,1000,0.48903272204243076,0.49154980222941386,0.48894282632146707
16
- 2,1500,0.49334771664868754,0.4964041711614527,0.4937072995325423
17
- 2,2000,0.4853469974829198,0.49658396260338006,0.4855267889248472
18
- 2,2500,0.4929881337648328,0.49379719525350596,0.49325782092772386
19
- 2,-1,0.4934376123696512,0.4962243797195254,0.4936174038115786
 
1
  epoch,steps,accuracy_cosinus,accuracy_manhattan,accuracy_euclidean
2
+ 0,200,0.6316073354908306,0.6332254584681769,0.6316073354908306
3
+ 0,200,0.5255303847536857,0.5258899676375405,0.5253505933117584
4
+ 0,400,0.6540812657317512,0.6537216828478964,0.6540812657317512
5
+ 0,200,0.6907587198849334,0.6819489392304926,0.6905789284430062
6
+ 0,400,0.6459906508450198,0.507371449119022,0.6456310679611651
7
+ 0,600,0.548363897878461,0.5456670262495505,0.548363897878461
8
+ 0,800,0.560230133045667,0.5593311758360302,0.560230133045667
9
+ 0,1000,0.531103919453434,0.5294857964760877,0.531103919453434
10
+ 0,1200,0.5213951815893564,0.5255303847536857,0.5213951815893564
11
+ 0,-1,0.5244516361021215,0.5255303847536857,0.5244516361021215
 
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:130ebd9b2818b76ec314cf2fed63db7c54e2192291a2fd3d1c4076d5c6d828cf
3
  size 437967672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fc2fcf3fd7c7c2c9253141c2e255c4468e0768f6e424dd820186aef7ad5b118
3
  size 437967672