# How to make 1. pretrained model: [epiCRealism](https://civitai.com/models/25694?modelVersionId=134065) + [hyper CFG lora 12steps](https://huggingface.co/ByteDance/Hyper-SD/blob/main/Hyper-SD15-12steps-CFG-lora.safetensors) -> merge with lora weight 0.3 2. model merged at step1 + lora model: [AnimateLCM_sd15_t2v_lora.safetensors](https://huggingface.co/wangfuyun/AnimateLCM/blob/main/AnimateLCM_sd15_t2v_lora.safetensors)-> merge with lora weight 0.8 ```python # Load the motion adapter adapter = MotionAdapter.from_pretrained("guoyww/animatediff-motion-adapter-v1-5-3", torch_dtype=torch.float16) # load SD 1.5 based finetuned model model_id = "/home/hyejin2/test/models/epiCRealism-hyper-LCM-8.safetensors" pipe = AnimateDiffVideoToVideoPipeline.from_single_file(model_id, motion_adapter=adapter, torch_dtype=torch.float16) pipe.save_pretrained("models/hello") ``` # How to use ```python model_id = "jstep750/animatediff_v2v" pipe = AnimateDiffVideoToVideoPipeline.from_pretrained(model_id, torch_dtype=torch.float16) # enable memory savings pipe.enable_vae_slicing() pipe.enable_model_cpu_offload() ```