Text Generation
Transformers
PyTorch
English
mistral
custom_code
text-generation-inference

any plans for mixtral 128k?

#17
by sirus - opened

Are there any plans to do this for mixtral? also with llama-cpp would 512GB of ram work for cpu inference or would I have to spend another $800 to move up to 1TB?

Thanks

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment