Danielus's picture

Danielus

danielus

AI & ML interests

None yet

Recent Activity

reacted to m-ric's post with šŸ”„ 11 days ago
š—£š—¼š˜š—²š—»š˜š—¶š—®š—¹ š—½š—®š—暝—®š—±š—¶š—“š—ŗ š˜€š—µš—¶š—³š˜ š—¶š—» š—Ÿš—Ÿš— š˜€: š—»š—²š˜„ š—½š—®š—½š—²š—æ š—Æš˜† š— š—²š˜š—® š—°š—¹š—®š—¶š—ŗš˜€ š˜š—µš—®š˜ š˜„š—² š—°š—®š—» š—“š—²š˜ š—暝—¶š—± š—¼š—³ š˜š—¼š—øš—²š—»š—¶š˜‡š—²š—暝˜€! šŸ„³ Current LLMs process text by first splitting it into tokens. They use a module named "tokenizer", that -spl-it-s- th-e- te-xt- in-to- arbitrary tokens depending on a fixed dictionnary. On the Hub you can find this dictionary in a model's files under tokenizer.json. āž”ļø This process is called BPE tokenization. It is suboptimal, everyone says it. It breaks text into predefined chunks that often fail to capture the nuance of language. But it has been a necessary evil in language models since their inception. šŸ’„ In Byte Latent Transformer (BLT), Meta researchers propose an elegant solution by eliminating tokenization entirely, working directly with raw bytes while maintaining efficiency through dynamic "patches." This had been tried before with different byte-level tokenizations, but it's the first time that an architecture of this type scales as well as BPE tokenization. And it could mean a real paradigm shift! šŸ‘šŸ‘ šŸ—ļø š—”š—暝—°š—µš—¶š˜š—²š—°š˜š˜‚š—暝—²: Instead of a lightweight tokenizer, BLT has a lightweight encoder that process raw bytes into patches. Then the patches are processed by the main heavy-duty transformers as we do normally (but for patches of bytes instead of tokens), before converting back to bytes. šŸ§© š——š˜†š—»š—®š—ŗš—¶š—° š—£š—®š˜š—°š—µš—¶š—»š—“: Instead of fixed tokens, BLT groups bytes based on their predictability (measured by entropy) - using more compute for complex sequences and efficiently handling simple ones. This allows efficient processing while maintaining byte-level understanding. I hope this breakthrough is confirmed and we can get rid of all the tokenizer stuff, it will make model handling easier! Read their paper here šŸ‘‰ https://dl.fbaipublicfiles.com/blt/BLT__Patches_Scale_Better_Than_Tokens.pdf
View all activity

Organizations

None yet

danielus's activity

New activity in infly/OpenCoder-8B-Instruct about 2 months ago

FIM task

4
#2 opened about 2 months ago by
danielus
New activity in bartowski/Qwen2.5-Coder-7B-Instruct-GGUF 3 months ago

FIM with Ollama

1
#2 opened 3 months ago by
danielus
New activity in lamhieu/ghost-8b-beta-8k 5 months ago
New activity in meta-llama/Llama-3.1-8B-Instruct 5 months ago

My alternative quantizations.

7
#16 opened 5 months ago by
ZeroWw

Garbage output ?

10
#30 opened 5 months ago by
danielus
New activity in DeepMount00/Qwen2-1.5B-Ita 6 months ago

Fine-tuning this model

2
#1 opened 6 months ago by
danielus
New activity in ExperimentLab/Mistral-ita-7b-Boost 8 months ago

LLama3 8B vs this model

2
#1 opened 8 months ago by
danielus
New activity in TroyDoesAI/Mermaid-Llama-3-8B 8 months ago

Long context

3
#2 opened 8 months ago by
danielus
New activity in TroyDoesAI/MermaidMixtral-3x7b 9 months ago

Best model

3
#1 opened 9 months ago by
danielus
New activity in TheBloke/StableBeluga2-70B-GGML over 1 year ago

Performance of quantified models

1
#3 opened over 1 year ago by
danielus