Training in progress, step 100, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 8668296
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:657bf78fae67ab7f060e521c89071af033d2ccaafb65ee8063a6a309b5d28838
|
3 |
size 8668296
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 17405562
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd88998b6ccade4db28e1d1fc48c4cbcfdb5f60c6ea942d1f805e4d77dac2e88
|
3 |
size 17405562
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 13990
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ecf02e2a13b83e9af6aadaf08c5e361bbd5737cddb097d92e6b793966187f367
|
3 |
size 13990
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4092f03770c9433b4ad0ca7df919cc90154d5482e807a1d8887f8b0bb6ea9041
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 5.905511811023622,
|
3 |
"best_model_checkpoint": "./exp/wft-test-model/checkpoint-80",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 10,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -762,6 +762,7 @@
|
|
762 |
},
|
763 |
{
|
764 |
"epoch": 4.05,
|
|
|
765 |
"grad_norm": 0.6178313493728638,
|
766 |
"learning_rate": 0.00019,
|
767 |
"loss": 0.1269,
|
@@ -769,6 +770,7 @@
|
|
769 |
},
|
770 |
{
|
771 |
"epoch": 4.06,
|
|
|
772 |
"grad_norm": 1.104316234588623,
|
773 |
"learning_rate": 0.00017999999999999998,
|
774 |
"loss": 0.2355,
|
@@ -776,6 +778,7 @@
|
|
776 |
},
|
777 |
{
|
778 |
"epoch": 4.07,
|
|
|
779 |
"grad_norm": 1.088135004043579,
|
780 |
"learning_rate": 0.00017,
|
781 |
"loss": 0.1947,
|
@@ -783,6 +786,7 @@
|
|
783 |
},
|
784 |
{
|
785 |
"epoch": 4.08,
|
|
|
786 |
"grad_norm": 0.6806421279907227,
|
787 |
"learning_rate": 0.00016,
|
788 |
"loss": 0.0923,
|
@@ -790,6 +794,7 @@
|
|
790 |
},
|
791 |
{
|
792 |
"epoch": 4.09,
|
|
|
793 |
"grad_norm": 0.6717160940170288,
|
794 |
"learning_rate": 0.00015,
|
795 |
"loss": 0.148,
|
@@ -797,6 +802,7 @@
|
|
797 |
},
|
798 |
{
|
799 |
"epoch": 4.1,
|
|
|
800 |
"grad_norm": 0.7663130760192871,
|
801 |
"learning_rate": 0.00014000000000000001,
|
802 |
"loss": 0.1192,
|
@@ -804,6 +810,7 @@
|
|
804 |
},
|
805 |
{
|
806 |
"epoch": 4.11,
|
|
|
807 |
"grad_norm": 1.271284818649292,
|
808 |
"learning_rate": 0.00013000000000000002,
|
809 |
"loss": 0.0887,
|
@@ -811,6 +818,7 @@
|
|
811 |
},
|
812 |
{
|
813 |
"epoch": 4.12,
|
|
|
814 |
"grad_norm": 0.6764338612556458,
|
815 |
"learning_rate": 0.00012,
|
816 |
"loss": 0.0812,
|
@@ -818,6 +826,7 @@
|
|
818 |
},
|
819 |
{
|
820 |
"epoch": 4.13,
|
|
|
821 |
"grad_norm": 0.9299374222755432,
|
822 |
"learning_rate": 0.00011,
|
823 |
"loss": 0.1574,
|
@@ -825,6 +834,7 @@
|
|
825 |
},
|
826 |
{
|
827 |
"epoch": 4.14,
|
|
|
828 |
"grad_norm": 0.9850154519081116,
|
829 |
"learning_rate": 0.0001,
|
830 |
"loss": 0.0678,
|
@@ -843,6 +853,90 @@
|
|
843 |
"eval_wer": 7.086614173228346,
|
844 |
"eval_wer_time": 0.007925033569335938,
|
845 |
"step": 90
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
846 |
}
|
847 |
],
|
848 |
"logging_steps": 1,
|
@@ -857,12 +951,12 @@
|
|
857 |
"should_evaluate": false,
|
858 |
"should_log": false,
|
859 |
"should_save": true,
|
860 |
-
"should_training_stop":
|
861 |
},
|
862 |
"attributes": {}
|
863 |
}
|
864 |
},
|
865 |
-
"total_flos":
|
866 |
"train_batch_size": 4,
|
867 |
"trial_name": null,
|
868 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 5.905511811023622,
|
3 |
"best_model_checkpoint": "./exp/wft-test-model/checkpoint-80",
|
4 |
+
"epoch": 5.05,
|
5 |
"eval_steps": 10,
|
6 |
+
"global_step": 100,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
762 |
},
|
763 |
{
|
764 |
"epoch": 4.05,
|
765 |
+
"eval_pred": null,
|
766 |
"grad_norm": 0.6178313493728638,
|
767 |
"learning_rate": 0.00019,
|
768 |
"loss": 0.1269,
|
|
|
770 |
},
|
771 |
{
|
772 |
"epoch": 4.06,
|
773 |
+
"eval_pred": null,
|
774 |
"grad_norm": 1.104316234588623,
|
775 |
"learning_rate": 0.00017999999999999998,
|
776 |
"loss": 0.2355,
|
|
|
778 |
},
|
779 |
{
|
780 |
"epoch": 4.07,
|
781 |
+
"eval_pred": null,
|
782 |
"grad_norm": 1.088135004043579,
|
783 |
"learning_rate": 0.00017,
|
784 |
"loss": 0.1947,
|
|
|
786 |
},
|
787 |
{
|
788 |
"epoch": 4.08,
|
789 |
+
"eval_pred": null,
|
790 |
"grad_norm": 0.6806421279907227,
|
791 |
"learning_rate": 0.00016,
|
792 |
"loss": 0.0923,
|
|
|
794 |
},
|
795 |
{
|
796 |
"epoch": 4.09,
|
797 |
+
"eval_pred": null,
|
798 |
"grad_norm": 0.6717160940170288,
|
799 |
"learning_rate": 0.00015,
|
800 |
"loss": 0.148,
|
|
|
802 |
},
|
803 |
{
|
804 |
"epoch": 4.1,
|
805 |
+
"eval_pred": null,
|
806 |
"grad_norm": 0.7663130760192871,
|
807 |
"learning_rate": 0.00014000000000000001,
|
808 |
"loss": 0.1192,
|
|
|
810 |
},
|
811 |
{
|
812 |
"epoch": 4.11,
|
813 |
+
"eval_pred": null,
|
814 |
"grad_norm": 1.271284818649292,
|
815 |
"learning_rate": 0.00013000000000000002,
|
816 |
"loss": 0.0887,
|
|
|
818 |
},
|
819 |
{
|
820 |
"epoch": 4.12,
|
821 |
+
"eval_pred": null,
|
822 |
"grad_norm": 0.6764338612556458,
|
823 |
"learning_rate": 0.00012,
|
824 |
"loss": 0.0812,
|
|
|
826 |
},
|
827 |
{
|
828 |
"epoch": 4.13,
|
829 |
+
"eval_pred": null,
|
830 |
"grad_norm": 0.9299374222755432,
|
831 |
"learning_rate": 0.00011,
|
832 |
"loss": 0.1574,
|
|
|
834 |
},
|
835 |
{
|
836 |
"epoch": 4.14,
|
837 |
+
"eval_pred": null,
|
838 |
"grad_norm": 0.9850154519081116,
|
839 |
"learning_rate": 0.0001,
|
840 |
"loss": 0.0678,
|
|
|
853 |
"eval_wer": 7.086614173228346,
|
854 |
"eval_wer_time": 0.007925033569335938,
|
855 |
"step": 90
|
856 |
+
},
|
857 |
+
{
|
858 |
+
"epoch": 4.15,
|
859 |
+
"grad_norm": 1.0147942304611206,
|
860 |
+
"learning_rate": 8.999999999999999e-05,
|
861 |
+
"loss": 0.0695,
|
862 |
+
"step": 91
|
863 |
+
},
|
864 |
+
{
|
865 |
+
"epoch": 4.16,
|
866 |
+
"grad_norm": 1.2335087060928345,
|
867 |
+
"learning_rate": 8e-05,
|
868 |
+
"loss": 0.1366,
|
869 |
+
"step": 92
|
870 |
+
},
|
871 |
+
{
|
872 |
+
"epoch": 4.17,
|
873 |
+
"grad_norm": 0.929300844669342,
|
874 |
+
"learning_rate": 7.000000000000001e-05,
|
875 |
+
"loss": 0.0959,
|
876 |
+
"step": 93
|
877 |
+
},
|
878 |
+
{
|
879 |
+
"epoch": 4.18,
|
880 |
+
"grad_norm": 0.5210400223731995,
|
881 |
+
"learning_rate": 6e-05,
|
882 |
+
"loss": 0.0852,
|
883 |
+
"step": 94
|
884 |
+
},
|
885 |
+
{
|
886 |
+
"epoch": 4.19,
|
887 |
+
"grad_norm": 0.8599880933761597,
|
888 |
+
"learning_rate": 5e-05,
|
889 |
+
"loss": 0.0508,
|
890 |
+
"step": 95
|
891 |
+
},
|
892 |
+
{
|
893 |
+
"epoch": 5.01,
|
894 |
+
"grad_norm": 0.38409674167633057,
|
895 |
+
"learning_rate": 4e-05,
|
896 |
+
"loss": 0.0801,
|
897 |
+
"step": 96
|
898 |
+
},
|
899 |
+
{
|
900 |
+
"epoch": 5.02,
|
901 |
+
"grad_norm": 0.5891416072845459,
|
902 |
+
"learning_rate": 3e-05,
|
903 |
+
"loss": 0.0762,
|
904 |
+
"step": 97
|
905 |
+
},
|
906 |
+
{
|
907 |
+
"epoch": 5.03,
|
908 |
+
"grad_norm": 0.658381998538971,
|
909 |
+
"learning_rate": 2e-05,
|
910 |
+
"loss": 0.0651,
|
911 |
+
"step": 98
|
912 |
+
},
|
913 |
+
{
|
914 |
+
"epoch": 5.04,
|
915 |
+
"grad_norm": 0.5465841293334961,
|
916 |
+
"learning_rate": 1e-05,
|
917 |
+
"loss": 0.1254,
|
918 |
+
"step": 99
|
919 |
+
},
|
920 |
+
{
|
921 |
+
"epoch": 5.05,
|
922 |
+
"grad_norm": 0.4337480664253235,
|
923 |
+
"learning_rate": 0.0,
|
924 |
+
"loss": 0.078,
|
925 |
+
"step": 100
|
926 |
+
},
|
927 |
+
{
|
928 |
+
"epoch": 5.05,
|
929 |
+
"eval_cer": 83.23863636363636,
|
930 |
+
"eval_cer_time": 0.003019094467163086,
|
931 |
+
"eval_decode_time": 0.5298569202423096,
|
932 |
+
"eval_loss": 0.118515744805336,
|
933 |
+
"eval_pred": "| i | Label | Prediction |\n| --- | --- | --- |\n| 0 | MISTER QUILTER IS THE APOSTLE OF THE MIDDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPEL | MISTER QUILTER IS THE APOSTLE OF THE MIDDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPELLLTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH-\"-\"-\"TH-\"-\"-\"-\"TH-\"-\"-\"-\"TH\"-TH\"-\"-\"-\"-THTH\"-\"-TH\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-=\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 1 | NOR IS MISTER QUILTER'S MANNER LESS INTERESTING THAN HIS MATTER | NOR IS MISTER QUILTER'S MANNER LESS INTERESTING THAN HIS MATTERTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH-\"-\"\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 2 | HE TELLS US THAT AT THIS FESTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEF LOOMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCCUR MOST READILY TO THE MIND | HE TELLS US THAT AT THIS FESTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEF LOOMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCCUR MOST READILY TO THE MINDTHTHTHTHTHTHTHTHTHTHTH%THTHTHTHTHTHTHTHTHTHTHTHTHTH%THTH%%%%%%%%THTH%%THTHTH%%%%%%%%%%%%%%%%%%%%%%%% |\n| 3 | HE HAS GRAVE DOUBTS WHETHER SIR FREDERICK LEIGHTON'S WORK IS REALLY GREEK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHACA | HE HAS GRAVE DOBTS WHETHER SIR FREDERICK LEIGHTON'S WORK IS REALLY GREEK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHACATHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH%%TH%% |\n| 4 | LINNELL'S PICTURES ARE A SORT OF UP GUARDS AND AT EM PAINTINGS AND MASON'S EXQUISITE IDYLLS ARE AS NATIONAL AS A JINGO POEM MISTER BIRKET FOSTER'S LANDSCAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARKER USED TO FLASH HIS TEETH AND MISTER JOHN COLLIER GIVES HIS SITTER A CHEERFUL SLAP ON THE BACK BEFORE HE SAYS LIKE A SHAMPOOER IN A TURKISH BATH NEXT MAN | LINNELL'S PICTURES ARE A SORT OF UP GUARDS AND AT EM PAINTINGS AND MASON'S EXQUISITE IDIDLLS ARE AS NATIONAL AS A JINGO POEM MISTER BIRKET FOSTER'S LANDSCAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARKER USED TO FLASH HIS TEETH AND MISTER JOHN COLLIER GIVES HIS SITTER A CHEERFUL SLAP ON THE BACK BEFORE HE SAYS LIKE A SHAMPOOER IN A TURKISH BATH NEXT MAN |\n| 5 | IT IS OBVIOUSLY UNNECESSARY FOR US TO POINT OUT HOW LUMINOUS THESE CRITICISMS ARE HOW DELICATE IN EXPRESSION | IT IS OBVIOUSLY UNNECESSARY FOR US TO POINT OUT HOW LUMINOUS THESE CRITICISMS ARE HOW DELICATE IN EXPRESSIONTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\" |\n| 6 | ON THE GENERAL PRINCIPLES OF ART MISTER QUILTER WRITES WITH EQUAL LUCIDITY | ON THE GENERAL PRINCIPLES OF ART MISTER QUILTER WRITES WITH EQUAL LUCIDITYTHTHTHTHTHTHTH \u266a \u266aTH \u266a%%==\"-\"-===\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 7 | PAINTING HE TELLS US IS OF A DIFFERENT QUALITY TO MATHEMATICS AND FINISH IN ART IS ADDING MORE FACT | PAINTING HE TELLS US IS OF A DIFFERENT QUALITY TO MATHEMATICS AND FINISH IN ART IS ADDING MORE FACT%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%\"-\"-%%%%%%%\"-\"-%%\"-%%%% |\n| 8 | AS FOR ETCHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGN | AS FOR ETCHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGNTHTHTHTHTHTHTHTHTHTH-\"-\"-\"-\"-\"-\"-\"-\" |\n| 9 | HE LAMENTS MOST BITTERLY THE DIVORCE THAT HAS BEEN MADE BETWEEN DECORATIVE ART AND WHAT WE USUALLY CALL PICTURES MAKES THE CUSTOMARY APPEAL TO THE LAST JUDGMENT AND REMINDS US THAT IN THE GREAT DAYS OF ART MICHAEL ANGELO WAS THE FURNISHING UPHOLSTERER | HE LAMENTS MOST BITTERLY THE DIVORCE THAT HAS BEEN MADE BETWEEN DECORATIVE ART AND WHAT WE USUALLY CALL PICTURES MAKES THE CUSTOMARY APPEAL TO THE LAST judGMENT AND REMINDS US THAT IN THE GREAT DAYS OF ART MIC ANGELO WAS THE FURNISHING UPHOLSTERER |\n",
|
934 |
+
"eval_runtime": 1.3526,
|
935 |
+
"eval_samples_per_second": 7.393,
|
936 |
+
"eval_steps_per_second": 1.479,
|
937 |
+
"eval_wer": 5.905511811023622,
|
938 |
+
"eval_wer_time": 0.004703044891357422,
|
939 |
+
"step": 100
|
940 |
}
|
941 |
],
|
942 |
"logging_steps": 1,
|
|
|
951 |
"should_evaluate": false,
|
952 |
"should_log": false,
|
953 |
"should_save": true,
|
954 |
+
"should_training_stop": true
|
955 |
},
|
956 |
"attributes": {}
|
957 |
}
|
958 |
},
|
959 |
+
"total_flos": 1.06772649984e+16,
|
960 |
"train_batch_size": 4,
|
961 |
"trial_name": null,
|
962 |
"trial_params": null
|