File size: 1,057 Bytes
d69a2f4
 
 
 
 
 
 
 
 
e1a3c98
bd2c5a0
e1a3c98
5d24c3e
bd2c5a0
03812d6
5d24c3e
bd2c5a0
5d24c3e
4ed7815
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
---
title: README
emoji: 🏢
colorFrom: purple
colorTo: pink
sdk: static
pinned: false
---

Welcome to the official Hugging Face organization for LLMQ. In this organization, you can find quantized models of LLM by cutting-edge quantization methods. 
In order to access models here, please select the suitable model for your personal use. 


We are dedicated to advancing the field of Artificial Intelligence with a focus on enhancing efficiency. Our primary research interests include quantiation, binarization, efficient learning, etc.
We are committed to innovating and developing cutting-edge techniques that make large language model (LLM) more accessible and sustainable, minimizing computational costs and maximizing performance. Our interdisciplinary approach leverages global expertise to push the boundaries of efficient AI technologies.

Recent Works:

[22.04.2024] How Good Are Low-bit Quantized LLaMA3 Models? An Empirical Study. Arxiv, 2024. [ArXiv](https://arxiv.org/abs/2404.14047) [GitHub](https://github.com/Macaronlin/LLaMA3-Quantization)