Post
1538
๐ข The LLaMA-3.1-8B distilled 8B version of the R1 DeepSeek AI is available besides the one based on Qwen
๐ Notebook for using it in reasoning over series of data ๐ง :
https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_deep_seek_7b_distill_llama3.ipynb
Loading using the pipeline API of the transformers library:
https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_llama.py
๐ก GPU Usage: 12.3 GB (FP16/FP32 mode) which is suitable for T4. (a 1.5 GB less than Qwen-distilled version)
๐ Perfomance: T4 instance: ~0.19 tokens/sec (FP32 mode) and (FP16 mode) ~0.22-0.30 tokens/sec. Is it should be that slow? ๐ค
Model name: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
โญ Framework: https://github.com/nicolay-r/bulk-chain
๐ Notebooks and models hub: https://github.com/nicolay-r/nlp-thirdgate
๐ Notebook for using it in reasoning over series of data ๐ง :
https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_deep_seek_7b_distill_llama3.ipynb
Loading using the pipeline API of the transformers library:
https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_llama.py
๐ก GPU Usage: 12.3 GB (FP16/FP32 mode) which is suitable for T4. (a 1.5 GB less than Qwen-distilled version)
๐ Perfomance: T4 instance: ~0.19 tokens/sec (FP32 mode) and (FP16 mode) ~0.22-0.30 tokens/sec. Is it should be that slow? ๐ค
Model name: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
โญ Framework: https://github.com/nicolay-r/bulk-chain
๐ Notebooks and models hub: https://github.com/nicolay-r/nlp-thirdgate