Update import.md
This commit is contained in:
parent
8ed51cac37
commit
176d0f7075
1 changed files with 1 additions and 12 deletions
|
@ -47,19 +47,13 @@ success
|
||||||
|
|
||||||
### Supported Quantizations
|
### Supported Quantizations
|
||||||
|
|
||||||
<details>
|
|
||||||
<summary>Legacy Quantization</summary>
|
|
||||||
|
|
||||||
- `Q4_0`
|
- `Q4_0`
|
||||||
- `Q4_1`
|
- `Q4_1`
|
||||||
- `Q5_0`
|
- `Q5_0`
|
||||||
- `Q5_1`
|
- `Q5_1`
|
||||||
- `Q8_0`
|
- `Q8_0`
|
||||||
|
|
||||||
</details>
|
#### K-means Quantizations
|
||||||
|
|
||||||
<details>
|
|
||||||
<summary>K-means Quantization</summary>`
|
|
||||||
|
|
||||||
- `Q3_K_S`
|
- `Q3_K_S`
|
||||||
- `Q3_K_M`
|
- `Q3_K_M`
|
||||||
|
@ -70,11 +64,6 @@ success
|
||||||
- `Q5_K_M`
|
- `Q5_K_M`
|
||||||
- `Q6_K`
|
- `Q6_K`
|
||||||
|
|
||||||
</details>
|
|
||||||
|
|
||||||
> [!NOTE]
|
|
||||||
> Activation-aware Weight Quantization (i.e. IQ) are not currently supported for automatic quantization however you can still import the quantized model into Ollama, see [Import GGUF](#import-gguf).
|
|
||||||
|
|
||||||
## Template Detection
|
## Template Detection
|
||||||
|
|
||||||
> [!NOTE]
|
> [!NOTE]
|
||||||
|
|
Loading…
Reference in a new issue