Commit History
add support for rpo_alpha (#1681)
c996881
unverified
re-enable DPO for tests in modal ci (#1374)
1f151c0
unverified
re-enable phi for tests in modal ci (#1373)
16d46b7
unverified
make sure the CI fails when pytest script fails (#1669)
fe650dd
unverified
Generalizing the chat_template prompt strategy (#1660) [skip ci]
cc11c6b
unverified
Fix llama3 chat_template (extra <|eot_id|> on last turn) (#1635)
7c2bf30
unverified
Add KTO support (#1640)
22ae21a
unverified
feat: Add LLaMA-3 instruct prompt strategies for fine-tuning (#1553)
50421c8
unverified
ADD: warning hub model (#1301)
601c08b
unverified
Add ORPO example and e2e test (#1572)
98c25e1
unverified
wrap prepared_ds_path in str() to avoid TypeError in fsspec package (#1548)
7477a53
unverified
ORPO Trainer replacement (#1551)
7d1d22f
unverified
fix broken linting (#1541)
c10563c
unverified
feat: validate sample packing requires flash_attention (#1465)
bf4cd67
unverified
Support loading datasets saved via save_to_disk (#1432)
e634118
unverified
make sure to capture non-null defaults from config validation (#1415)
601b77b
unverified
fix(dataset): normalize tokenizer config and change hash from tokenizer class to tokenizer path (#1298)
ff939d8
unverified
strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified
Feat: Add sharegpt multirole (#1137)
40a88e8
unverified
ORPO (#1419)
2ea70eb
unverified
Train parameters exclusively in specific ranges (#1390)
05bcc9e
unverified
Add Glaive conversation format support (#1365)
b7d8a7d
unverified
plain input/output prompt strategy w/o chat templates (#1346)
4d09b42
unverified
run tests again on Modal (#1289) [skip ci]
0001862
unverified
fix for protected model_ namespace w pydantic (#1345)
6b3b271
unverified
more fixes 20240228 (#1342) [skip ci]
0f985e1
unverified
Pydantic 2.x cfg (#1239)
cc3cebf
unverified
make mlflow optional (#1317)
5894f0e
unverified
Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (#1273)
8430db2
unverified
Pretrain transforms (#1261)
c7cf381
unverified
relora: magnitude pruning of the optimizer (#1245)
8c2e05a
unverified
support for true batches with multipack (#1230)
00568c1
unverified
Support for additional_special_tokens (#1221) [skip ci]
25e037f
unverified
Peft lotfq (#1222)
4cb7900
unverified
ADD: warning if hub_model_id ist set but not any save strategy (#1202)
af29d81
unverified
Feat/chatml add system message (#1117)
98b4762
unverified
Phi2 multipack (#1173)
814aee6
unverified
Feat(test): Add tests for alpaca chatml prompt tokenizer (#1088)
5439707
unverified
Falcon embeddings (#1149) [skip docker]
e799e08
unverified
Deprecate max packed sequence len (#1141)
2ce5c0d
unverified
Multipack simplify for Mixtral (#1142)
6910e6a
unverified
Add `layers_to_transform` for `lora_config` (#1118)
8487b97
unverified
xzuyn
commited on
Enable or disable bf16 support based on availability (#1116)
0865613
unverified
Simon Hällqvist
commited on