results

This model is a fine-tuned version of facebook/detr-resnet-50 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.0368

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 32
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • training_steps: 2000

Training results

Training Loss Epoch Step Validation Loss
3.4991 0.3137 100 3.6244
3.1728 0.3451 110 3.4319
2.7857 0.3765 120 3.2574
3.0606 0.4078 130 3.1484
2.6704 0.4392 140 3.0390
2.7332 0.4706 150 2.9956
2.8436 0.5020 160 2.9110
2.8464 0.5333 170 2.8551
2.1192 0.5647 180 2.8163
2.6557 0.5961 190 2.8145
2.3224 0.6275 200 2.7858
2.6007 0.6588 210 2.7064
2.6117 0.6902 220 2.6602
2.4549 0.7216 230 2.6368
2.5487 0.7529 240 2.6029
2.6048 0.7843 250 2.5573
2.0348 0.8157 260 2.5203
2.4741 0.8471 270 2.4935
2.5855 0.8784 280 2.4731
2.1076 0.9098 290 2.4283
2.3073 0.9412 300 2.3896
2.214 0.9725 310 2.3919
2.2078 1.0039 320 2.3343
2.2391 1.0353 330 2.2970
2.3607 1.0667 340 2.2921
2.0244 1.0980 350 2.2751
2.251 1.1294 360 2.2713
2.1133 1.1608 370 2.2701
2.124 1.1922 380 2.2618
2.1989 1.2235 390 2.2429
2.2315 1.2549 400 2.2463
2.2398 1.2863 410 2.2386
2.261 1.3176 420 2.2360
2.2144 1.3490 430 2.2427
2.3344 1.3804 440 2.2452
2.0412 1.4118 450 2.2092
2.0854 1.4431 460 2.2197
2.1636 1.4745 470 2.1830
1.7776 1.5059 480 2.1904
2.1118 1.5373 490 2.2194
2.1203 1.5686 500 2.1978
2.2468 1.6 510 2.1968
2.2992 1.6314 520 2.1963
2.2596 1.6627 530 2.1816
2.1836 1.6941 540 2.1800
2.2672 1.7255 550 2.1679
2.0702 1.7569 560 2.1607
2.5606 1.7882 570 2.1568
2.1392 1.8196 580 2.1578
1.9255 1.8510 590 2.1799
2.0995 1.8824 600 2.1995
2.1153 1.9137 610 2.1741
2.2068 1.9451 620 2.1638
1.8698 1.9765 630 2.1819
1.8849 2.0078 640 2.1807
2.0291 2.0392 650 2.1636
2.2092 2.0706 660 2.1356
2.1117 2.1020 670 2.1682
1.8318 2.1333 680 2.1719
1.9884 2.1647 690 2.2114
2.1933 2.1961 700 2.1526
2.2953 2.2275 710 2.1525
2.2841 2.2588 720 2.1417
1.9865 2.2902 730 2.1399
1.9193 2.3216 740 2.1313
1.8882 2.3529 750 2.1362
1.8967 2.3843 760 2.1454
1.9424 2.4157 770 2.1356
1.8531 2.4471 780 2.1340
1.9435 2.4784 790 2.1413
2.0455 2.5098 800 2.1558
1.9384 2.5412 810 2.1519
2.0826 2.5725 820 2.1381
2.0008 2.6039 830 2.1136
1.922 2.6353 840 2.1160
1.9567 2.6667 850 2.0991
2.2798 2.6980 860 2.0998
2.4014 2.7294 870 2.0922
2.3427 2.7608 880 2.0976
2.2701 2.7922 890 2.0823
2.1405 2.8235 900 2.1009
1.9259 2.8549 910 2.1075
2.0055 2.8863 920 2.1041
1.9902 2.9176 930 2.0854
1.9821 2.9490 940 2.1107
2.0292 2.9804 950 2.0901
1.9811 3.0118 960 2.1227
2.2674 3.0431 970 2.0934
2.0632 3.0745 980 2.0935
2.1232 3.1059 990 2.0843
2.0056 3.1373 1000 2.0891
2.0188 3.1686 1010 2.0811
2.0898 3.2 1020 2.0848
2.1809 3.2314 1030 2.0883
2.1636 3.2627 1040 2.0931
1.9941 3.2941 1050 2.0894
1.9761 3.3255 1060 2.0957
1.9908 3.3569 1070 2.0715
2.0806 3.3882 1080 2.0774
1.9419 3.4196 1090 2.0713
1.8643 3.4510 1100 2.0654
1.969 3.4824 1110 2.0636
2.0104 3.5137 1120 2.0710
1.6745 3.5451 1130 2.0551
2.047 3.5765 1140 2.0598
2.1289 3.6078 1150 2.0426
2.1158 3.6392 1160 2.0525
1.8543 3.6706 1170 2.0515
2.0206 3.7020 1180 2.0508
2.1992 3.7333 1190 2.0485
1.6875 3.7647 1200 2.0558
1.8452 3.7961 1210 2.0543
2.2061 3.8275 1220 2.0594
2.0418 3.8588 1230 2.0652
2.0411 3.8902 1240 2.0679
2.0835 3.9216 1250 2.0731
1.9003 3.9529 1260 2.0574
1.7881 3.9843 1270 2.0777
2.1354 4.0157 1280 2.0630
1.8935 4.0471 1290 2.0607
2.1067 4.0784 1300 2.0576
1.8225 4.1098 1310 2.0767
1.8132 4.1412 1320 2.0507
1.985 4.1725 1330 2.0669
2.112 4.2039 1340 2.0836
1.7993 4.2353 1350 2.0718
1.9784 4.2667 1360 2.0676
2.1628 4.2980 1370 2.0525
1.876 4.3294 1380 2.0615
2.0081 4.3608 1390 2.0736
1.8642 4.3922 1400 2.0565
1.9308 4.4235 1410 2.0608
2.2296 4.4549 1420 2.0553
2.0166 4.4863 1430 2.0575
2.0422 4.5176 1440 2.0543
1.8729 4.5490 1450 2.0552
2.0323 4.5804 1460 2.0656
1.9935 4.6118 1470 2.0794
1.8534 4.6431 1480 2.0685
1.8363 4.6745 1490 2.0581
1.9679 4.7059 1500 2.0353
1.8585 4.7373 1510 2.0334
1.9772 4.7686 1520 2.0420
1.8753 4.8 1530 2.0427
1.8911 4.8314 1540 2.0499
2.0614 4.8627 1550 2.0481
2.1184 4.8941 1560 2.0481
1.9504 4.9255 1570 2.0541
2.1337 4.9569 1580 2.0480
2.4391 4.9882 1590 2.0416
1.72 5.0196 1600 2.0412
2.0808 5.0510 1610 2.0458
1.8639 5.0824 1620 2.0438
1.9462 5.1137 1630 2.0428
2.0055 5.1451 1640 2.0366
2.0345 5.1765 1650 2.0644
1.9321 5.2078 1660 2.0454
1.8705 5.2392 1670 2.0394
2.0345 5.2706 1680 2.0475
1.9992 5.3020 1690 2.0567
2.2208 5.3333 1700 2.0558
1.8253 5.3647 1710 2.0413
2.0765 5.3961 1720 2.0319
2.2315 5.4275 1730 2.0360
2.2432 5.4588 1740 2.0436
2.0666 5.4902 1750 2.0451
2.0603 5.5216 1760 2.0296
1.6625 5.5529 1770 2.0513
2.0946 5.5843 1780 2.0306
1.9464 5.6157 1790 2.0315
2.0183 5.6471 1800 2.0276
2.0794 5.6784 1810 2.0512
2.0289 5.7098 1820 2.0369
2.1014 5.7412 1830 2.0520
1.9159 5.7725 1840 2.0491
2.2446 5.8039 1850 2.0508
1.9383 5.8353 1860 2.0327
2.0132 5.8667 1870 2.0161
2.2234 5.8980 1880 2.0406
2.2556 5.9294 1890 2.0365
2.2061 5.9608 1900 2.0314
1.7465 5.9922 1910 2.0543
1.9388 6.0235 1920 2.0525
1.9223 6.0549 1930 2.0325
1.9386 6.0863 1940 2.0282
1.9171 6.1176 1950 2.0462
1.9319 6.1490 1960 2.0369
1.7689 6.1804 1970 2.0364
2.0063 6.2118 1980 2.0388
2.1053 6.2431 1990 2.0346
2.1074 6.2745 2000 2.0368

Framework versions

  • Transformers 4.41.2
  • Pytorch 2.3.1
  • Datasets 2.19.2
  • Tokenizers 0.19.1
Downloads last month
17
Safetensors
Model size
41.6M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for amaye15/results

Finetuned
(456)
this model