Flash attention for udop large

#8
by Vasanth2002 - opened

I am trying to increase the inference speed of UDOP but nothing works. I want to use flashattention for udop but it is not supported

Feel free to open an issue on Github regarding adding support

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment