Robert Shaw
robertgshaw2
AI & ML interests
None yet
Recent Activity
updated
a model
about 19 hours ago
nm-testing/Mixtral-8x7B-Instruct-v0.1-FP8-Dynamic
published
a model
about 19 hours ago
nm-testing/Mixtral-8x7B-Instruct-v0.1-FP8-Dynamic
new activity
about 1 month ago
neuralmagic/DeepSeek-R1-Distill-Llama-70B-FP8-dynamic:Update tokenizer_config.json
Organizations
robertgshaw2's activity
Update tokenizer_config.json
5
#3 opened about 1 month ago
by
erichartford

What is an actorder group and what are the advantages of running this in vLLM?
4
#1 opened 3 months ago
by
nickandbro
Can I apply a LoRA?
2
#1 opened 3 months ago
by
RonanMcGovern

Nice model, any info on scripts used to quantize?
1
#1 opened 3 months ago
by
RonanMcGovern

How to download the model with transformer library
5
#6 opened 5 months ago
by
Rick10
Update README.md
3
#25 opened 5 months ago
by
robertgshaw2
Issue running on vLLM using FP8
2
#3 opened 6 months ago
by
ffleandro
vllm says the requested model does not exist
2
#1 opened 7 months ago
by
shivams101
Storage format differs from other w4a16 models
2
#2 opened 7 months ago
by
timdettmers

Model weights are not loaded
4
#3 opened 7 months ago
by
MarvelousMouse

Can not be inferenced with vllm openai server
1
#1 opened 8 months ago
by
jjqsdq
Code example request with vllm
2
#1 opened 9 months ago
by
ShiningJazz
4bit quantisation does not reduce vram usage.
1
#2 opened 9 months ago
by
fu-man
How to run Meta-Llama-3-70B-Instruct-FP8 using several devices?
5
#3 opened 9 months ago
by
Fertel
Reproduction
2
#792 opened 9 months ago
by
robertgshaw2
Fails to run with nm-vllm
1
#1 opened 11 months ago
by
clintonruairi
Update chart template
#2 opened 12 months ago
by
robertgshaw2
Update chart template
#2 opened 12 months ago
by
robertgshaw2