Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention Paper โข 2404.07143 โข Published Apr 10, 2024 โข 111
Star Attention: Efficient LLM Inference over Long Sequences Paper โข 2411.17116 โข Published Nov 26, 2024 โข 56
Qwen2 Collection Qwen2 language models, including pretrained and instruction-tuned models of 5 sizes, including 0.5B, 1.5B, 7B, 57B-A14B, and 72B. โข 39 items โข Updated 27 days ago โข 368