Spaces:
Sleeping
Sleeping
File size: 2,714 Bytes
15d6c34 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
## INSTALLATION
if you've noticed your python3 bin doens't point to your conda env when using --prefix to point to your scratch dir, then you need to do the following:
* conda config --set always_copy True
* conda config --show | grep always_copy
now continue as normal:
* conda create --prefix <your-scratch-path>/MotionDiffuse/env python=3.7
* conda activate <your-scratch-path>/MotionDiffuse/env
* double check your GCC is 5+ by running `gcc --version`; if not, do module load gcc/5.4.0
* module load cuda/10.1 # you must run these icuda commands before installing torch otherwise it will say version not found!!
* module load cudnn/v7.6.5.32-prod-cuda-10.1
* conda install pytorch=1.7.1 torchvision=0.8.2 cudatoolkit=10.1 -c pytorch
* python3 -m pip install "mmcv-full>=1.3.17,<=1.5.3" -f https://download.openmmlab.com/mmcv/dist/cu101/torch1.7.1/index.html
* python3 -m pip install -r requirements.txt
* python3 -m pip install --upgrade protobuf==3.20.0
fyi there is an annoying warning in the logs (https://stackoverflow.com/questions/57381430/synonym-of-type-is-deprecated-in-a-future-version-of-numpy-it-will-be-underst) that can be silenced by downgrading numpy:to 1.16.4 BUT this is incompatible with the other package versions, so don't do it
fyi:
(/work3/s222376/MotionDiffuseNew) s222376@n-62-20-1 /work3/s222376/MotionDiffuse/text2motion (train_baseline)$ module list
Currently Loaded Modulefiles:
1) latex/TeXLive19(default) 3) cudnn/v7.6.5.32-prod-cuda-10.1 5) gcc/5.4.0
2) cuda/10.1 4) binutils/2.29(default) <aL>
## TRAINING
* download KIT-ML data from <> and put the zip for it in text2motion/data/
* cd text2motion/data && unzip KIT-ML-20231122T121619Z-001.zip
* cd KIT-ML && unrar x new_joint_vecs.rar
* unrar x new_joints.rar
* unrar x texts.rar
* dirs should look like
```
text2motion/data/KIT-ML
βββ new_joint_vecs
βΒ Β ββοΏ½
βββ new_joints
βΒ Β ββοΏ½
βββ texts
ββοΏ½
--all.txt
--<etc>
```
* voltash (dtu hpc command to go to interactive gpu node)
* make train
* verify above works without errors and then kill training because you're on interactive gpu, you will likely run out of memory anyway (can decrease --batchsize but then it's slow)
* to do full training, edit jobscript.sh to use your email and submit job via "make queue"
## INFERENCE with pretrained model
* download...checkpoints?? idk look at their README.md
## Changes I made
* ignore standardization
* tokens are [] empty...
* reusing kit_chain thing lol
* only training on one sequence from grab
TO KEEP IN MIND:
* they specify best way to train in readme somewhere -- follow this when doing real training!
* need to add the emotion text to the caption!! |