consolidated vs model safetensors - what's the difference?

#9
by jukofyork - opened

Just looking to download this and noticed there are two sets of safetensors files? I assume it's the "model" ones that need to be downloaded?

The model files are added in PR #1 so those are the ones I'm grabbing - I'm assuming they're more correct because they're more recent? ¯_(ツ)_/¯

why are there two copies of the model here. this sucks. what a waste of bandwidth and time...

hugginface-cli download ..... --exclude="*consolidated*" to skip over the consolidated files. I don't think the companies realize not everyone work/live with 10G-100G uplinks. =)

You have to first know there are multiple copies of the model in the repo
By the time you know you've already finished downloading the wrong one and started downloading the right one

Just wish the big guys would look around a little bit with a curiousity to fit in and learn our ways. We have good reasons for our conventions

It seems that the consolidated*-safetentors are the correct ones, according to the Usage in Model card.

Hey all, I opened the initial PR to offer the transformers' implementation out of the box with the repo.

To me it's quite similar to other repos where you might have multiple quants; you don't download all the quants, same as here you wouldn't download both implementations.

How do you get the files locally? Cloning the repo is a way but really isn't the best way to do so IMO: both mistral-inference and transformers will only download the files that they need, and in case you want to use a CLI you can use what @Qubitium mentions with huggingface-cli download and the --include and --excluse flags. Using this in conjunction with HF_HUB_ENABLE_HF_TRANSFER will be much much faster than using git clone to obtain the files.

Maybe for future repos we could include download options in the model card for both implementations, so that it's easier to see how to get about to downloading each of the two rather than both (which no-one wants to do, agree with all of you).

Mistral AI_ org

Just wish the big guys would look around a little bit with a curiousity to fit in and learn our ways. We have good reasons for our conventions

I understand the issue, but in this case what do you believe would be the best approach? As lysandre mentions this is to be able to offer the transformers implementation. Is there a recommended way you believe it would be better to go with?

Use two repos. (Just like everyone else does)

@ehartford why are you complaining? (again)

You want them to create 2 seperate repos just because you won't take 1 second out of your day to check what files you're downloading?

By the time you know you've already finished downloading

If your internet is that fast, why do you care?

fit in and learn our ways

lol

You made finetunes (which aren't that great) and now you act like you own the place. Get over yourself.

I fully support Mistral on this one. Who cares about the environment and HF bandwidth? You rock Mistral!!!

deleted

While we all appreciate what mistral does, cluttering things up like that is silly.

Just learned a magic trick: You can download Mistral Large twice by forgetting to add --exclude="*consolidated*". It's a neat way to waste bandwidth, time, money, and disk space!

Save your breath. They don't care.

deleted

Just learned a magic trick: You can download Mistral Large twice by forgetting to add --exclude="*consolidated*". It's a neat way to waste bandwidth, time, money, and disk space!

lol

Sign up or log in to comment