|
344 | 344 | # LLAMA_FTYPE_MOSTLY_IQ4_XS = 30, // except 1d tensors |
345 | 345 | # LLAMA_FTYPE_MOSTLY_IQ1_M = 31, // except 1d tensors |
346 | 346 | # LLAMA_FTYPE_MOSTLY_BF16 = 32, // except 1d tensors |
347 | | -# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // except 1d tensors |
348 | | -# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // except 1d tensors |
349 | | -# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // except 1d tensors |
| 347 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // removed from gguf files, use Q4_0 and runtime repack |
| 348 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // removed from gguf files, use Q4_0 and runtime repack |
| 349 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // removed from gguf files, use Q4_0 and runtime repack |
350 | 350 | # LLAMA_FTYPE_MOSTLY_TQ1_0 = 36, // except 1d tensors |
351 | 351 | # LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors |
352 | 352 | # |
|
382 | 382 | LLAMA_FTYPE_MOSTLY_IQ4_XS = 30 |
383 | 383 | LLAMA_FTYPE_MOSTLY_IQ1_M = 31 |
384 | 384 | LLAMA_FTYPE_MOSTLY_BF16 = 32 |
385 | | -LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33 |
386 | | -LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34 |
387 | | -LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35 |
| 385 | +# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33 |
| 386 | +# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34 |
| 387 | +# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35 |
388 | 388 | LLAMA_FTYPE_MOSTLY_TQ1_0 = 36 |
389 | 389 | LLAMA_FTYPE_MOSTLY_TQ2_0 = 37 |
390 | 390 | LLAMA_FTYPE_GUESSED = 1024 |
|
0 commit comments