Haotong Qin commited on
Commit
bd2c5a0
·
verified ·
1 Parent(s): ec45061

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -10
README.md CHANGED
@@ -7,18 +7,13 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- The LLMQ Family
11
-
12
- From ETH Zuich, Beihang University, The University of Hong Kong
13
-
14
  Welcome to the official Hugging Face organization for LLMQ. In this organization, you can find quantized models of LLM by cutting-edge quantization methods.
15
- In order to access models here, please selected the suitable model for your personal use.
16
-
17
 
18
- Team LLMQ is dedicated to advancing the field of Artificial Intelligence with a focus on enhancing efficiency. Our primary research interests include quantiation, binarization, efficient learning, etc.
19
- We are committed to innovate and develop cutting-edge techniques that make AI more accessible and sustainable, minimizing computational costs and maximizing performance. Our interdisciplinary approach leverages global expertise to push the boundaries of efficient AI technologies.
20
 
 
 
21
 
22
- Our team is organized by Dr.Qin, who is a Postdoctoral Researcher at the Center for Project-Based Learning (PBL), ETH Zürich, Switzerland.
23
 
24
- Contact Information: qinhaotong@gmail.com
 
7
  pinned: false
8
  ---
9
 
 
 
 
 
10
  Welcome to the official Hugging Face organization for LLMQ. In this organization, you can find quantized models of LLM by cutting-edge quantization methods.
11
+ In order to access models here, please select the suitable model for your personal use.
 
12
 
 
 
13
 
14
+ We are dedicated to advancing the field of Artificial Intelligence with a focus on enhancing efficiency. Our primary research interests include quantiation, binarization, efficient learning, etc.
15
+ We are committed to innovating and developing cutting-edge techniques that make AI more accessible and sustainable, minimizing computational costs and maximizing performance. Our interdisciplinary approach leverages global expertise to push the boundaries of efficient AI technologies.
16
 
17
+ Recent Works:
18
 
19
+ [22.04.2024] How Good Are Low-bit Quantized LLaMA3 Models? An Empirical Study. Arxiv, 2024. [ArXiv]() [GitHub]()