Run with full 128k context in 24G vram

#3
by meigami - opened

its a amazing model but need too much vram, I cannot run with full 128k context in 24G vram
Do you have any plans to add lighter versions? @TheBloke @LoneStriker

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment