Spaces:
Running
on
Zero
Running
on
Zero
File size: 1,685 Bytes
841bef5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
# --------------------------------------------------------
# InternVL
# Copyright (c) 2024 OpenGVLab
# Licensed under The MIT License [see LICENSE for details]
# --------------------------------------------------------
from .internlm2_packed_training_patch import replace_internlm2_attention_class
from .internvit_liger_monkey_patch import apply_liger_kernel_to_internvit
from .llama2_flash_attn_monkey_patch import replace_llama2_attn_with_flash_attn
from .llama_flash_attn_monkey_patch import replace_llama_attn_with_flash_attn
from .llama_packed_training_patch import replace_llama_attention_class
from .llama_rmsnorm_monkey_patch import \
replace_llama_rmsnorm_with_fused_rmsnorm
from .pad_data_collator import (concat_pad_data_collator,
dpo_concat_pad_data_collator,
pad_data_collator)
from .phi3_packed_training_patch import replace_phi3_attention_class
from .qwen2_packed_training_patch import replace_qwen2_attention_class
from .train_dataloader_patch import replace_train_dataloader
from .train_sampler_patch import replace_train_sampler
__all__ = ['replace_llama_attn_with_flash_attn',
'replace_llama_rmsnorm_with_fused_rmsnorm',
'replace_llama2_attn_with_flash_attn',
'replace_train_sampler',
'replace_train_dataloader',
'replace_internlm2_attention_class',
'replace_qwen2_attention_class',
'replace_phi3_attention_class',
'replace_llama_attention_class',
'pad_data_collator',
'dpo_concat_pad_data_collator',
'concat_pad_data_collator',
'apply_liger_kernel_to_internvit']
|