--- license: apache-2.0 datasets: - Open-Orca/SlimOrca language: - en --- ### TeeZee/GALAXY-XB-v1.03 ### Experiment, can DUS be taken one or more steps further? ### Technical notes: - model v03 finetuned on 50k entries from SlimOrca dataset - 12 layers removed from both models, 4 more than in original paper but its 1/4 of all layers(48) as per original paper. - base version of upstage/SOLAR-10.7B-v1.0 used for merge ### To evaluate - model performance after finetuning, did it recover initial performance loss after merge?