SandLogicTechnologies nielsr HF Staff commited on
Commit
0ff8055
·
verified ·
1 Parent(s): e7176b5

Add link to paper (#2)

Browse files

- Add link to paper (c4fd121230b81373c16eba1cdd23be19e7acdf15)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -1,17 +1,18 @@
1
  ---
2
- pipeline_tag: text-generation
3
- library_name: transformers
4
- license: apache-2.0 # Assuming Apache 2.0 license, adjust if different.
5
  base_model:
6
  - Qwen/Qwen3-8B
7
  - Qwen/Qwen3-0.6B
8
  - Qwen/Qwen3-4B
9
  - Qwen/Qwen3-1.7B
 
 
 
10
  tags:
11
  - Light weight
12
  - Agentic
13
  - Conversational
14
  ---
 
15
  # Qwen3 Quantized Models – Lexicons Edition
16
 
17
  This repository provides quantized versions of the **Qwen3** language models, optimized for efficient deployment on edge devices and low-resource environments. The following models have been added to our **Lexicons** Model Zoo:
@@ -25,7 +26,7 @@ This repository provides quantized versions of the **Qwen3** language models, op
25
 
26
  ## Model Overview
27
 
28
- **Qwen3** is the latest open-source LLM series developed by Alibaba Group. Released on **April 28, 2025**, the models were trained on **36 trillion tokens** across **119 languages and dialects**. Qwen3 models are instruction-tuned and support long context windows and multilingual capabilities. This model is described in [An Empirical Study of Qwen3 Quantization](https://arxiv.org/abs/2505.02214).
29
 
30
  The quantized versions provided here use **4-bit Q4_K_M** precision ensuring high performance at a fraction of the memory and compute cost. These models are ideal for real-time inference, chatbots, and on-device applications.
31
 
@@ -38,7 +39,6 @@ The quantized versions provided here use **4-bit Q4_K_M** precision ensuring hig
38
  - **Instruction-Tuned**: Fine-tuned to follow user instructions effectively.
39
  - **Scalable Sizes**: Choose from 0.6B to 8B parameter models based on your use case.
40
 
41
-
42
  ---
43
 
44
  ## Available Quantized Versions
 
1
  ---
 
 
 
2
  base_model:
3
  - Qwen/Qwen3-8B
4
  - Qwen/Qwen3-0.6B
5
  - Qwen/Qwen3-4B
6
  - Qwen/Qwen3-1.7B
7
+ library_name: transformers
8
+ license: apache-2.0
9
+ pipeline_tag: text-generation
10
  tags:
11
  - Light weight
12
  - Agentic
13
  - Conversational
14
  ---
15
+
16
  # Qwen3 Quantized Models – Lexicons Edition
17
 
18
  This repository provides quantized versions of the **Qwen3** language models, optimized for efficient deployment on edge devices and low-resource environments. The following models have been added to our **Lexicons** Model Zoo:
 
26
 
27
  ## Model Overview
28
 
29
+ **Qwen3** is the latest open-source LLM series developed by Alibaba Group. Released on **April 28, 2025**, the models were trained on **36 trillion tokens** across **119 languages and dialects**. Qwen3 models are instruction-tuned and support long context windows and multilingual capabilities. This model is described in [An Empirical Study of Qwen3 Quantization](https://huggingface.co/papers/2505.02214).
30
 
31
  The quantized versions provided here use **4-bit Q4_K_M** precision ensuring high performance at a fraction of the memory and compute cost. These models are ideal for real-time inference, chatbots, and on-device applications.
32
 
 
39
  - **Instruction-Tuned**: Fine-tuned to follow user instructions effectively.
40
  - **Scalable Sizes**: Choose from 0.6B to 8B parameter models based on your use case.
41
 
 
42
  ---
43
 
44
  ## Available Quantized Versions