Kensho: a luminous awakening where the veil of illusion dissolves, revealing the boundless truth of our interconnected essence, inviting us into a dance with the infinite.

image/png

By Fernando, Eric and David

Discord Discord: https://discord.gg/cognitivecomputations

This is a hack around pytorch + huggingface Transformers library to make the original Dolphin Phi-2 to behave in a way inspired by the Meta's paper "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases" [ https://arxiv.org/abs/2402.14905 ]

One of the key ideas is that it works as if it was like "an online passthrough", by applying a loop on a module SuperClass, that groups layers, in a such way they get their forward method repeated in a loop. So, in theory, you can observe more intelligence in the same way MegaDolphin 120b, Professor 155b, Venus120b and other huge models, but use way less vRAM, because instead of cloning the weights, we share them in the vRAM.

And actually, this concept could be also used to enable the training of way more efficient models.

We hope the community enjoy it and make good use of it.

It won't work out of the box in the other models. Their "modeling files" should be changed accordingly to achieve the same effect.

Downloads last month
30
Safetensors
Model size
2.78B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for cognitivecomputations/dolphin-phi-2-kensho

Finetunes
7 models