sync : llama.cpp (ggml/0)

ggml-ci
This commit is contained in:
Georgi Gerganov
2024-02-21 16:19:39 +02:00
parent 208de95ac7
commit ce411498f6
8 changed files with 626 additions and 3 deletions

View File

@ -66,6 +66,7 @@ bool ggml_common_quantize_0(
case GGML_FTYPE_MOSTLY_IQ2_XS:
case GGML_FTYPE_MOSTLY_IQ3_XXS:
case GGML_FTYPE_MOSTLY_IQ1_S:
case GGML_FTYPE_MOSTLY_IQ4_NL:
{
fprintf(stderr, "%s: invalid model type %d\n", __func__, ftype);
return false;
@ -199,6 +200,7 @@ bool ggml_common_quantize_0(
case GGML_TYPE_IQ2_XS:
case GGML_TYPE_IQ3_XXS:
case GGML_TYPE_IQ1_S:
case GGML_TYPE_IQ4_NL:
case GGML_TYPE_COUNT:
{
fprintf(stderr, "%s: unsupported quantization type %d (%s)\n", __func__, ttype, ggml_type_name((ggml_type) ttype));