mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 13:20:12 +06:00
Support loading Gemma3 QAT GGUF models (#37649)
* fix gemma3 qat gguf support Signed-off-by: isotr0py <2037008807@qq.com> * update test Signed-off-by: isotr0py <2037008807@qq.com> * make ruff happy Signed-off-by: isotr0py <2037008807@qq.com> --------- Signed-off-by: isotr0py <2037008807@qq.com> Co-authored-by: Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
This commit is contained in:
parent
7eb1107cc2
commit
c69e23455d
@ -258,6 +258,8 @@ TENSOR_PROCESSORS = {
|
|||||||
|
|
||||||
|
|
||||||
def read_field(reader, field):
|
def read_field(reader, field):
|
||||||
|
if field not in reader.fields:
|
||||||
|
return []
|
||||||
value = reader.fields[field]
|
value = reader.fields[field]
|
||||||
return [_gguf_parse_value(value.parts[_data_index], value.types) for _data_index in value.data]
|
return [_gguf_parse_value(value.parts[_data_index], value.types) for _data_index in value.data]
|
||||||
|
|
||||||
@ -369,6 +371,7 @@ def load_gguf_checkpoint(gguf_checkpoint_path, return_tensors=False, model_to_lo
|
|||||||
parsed_parameters = {k: {} for k in GGUF_TO_TRANSFORMERS_MAPPING}
|
parsed_parameters = {k: {} for k in GGUF_TO_TRANSFORMERS_MAPPING}
|
||||||
|
|
||||||
architecture = read_field(reader, "general.architecture")[0]
|
architecture = read_field(reader, "general.architecture")[0]
|
||||||
|
# NOTE: Some GGUF checkpoints may miss `general.name` field in metadata
|
||||||
model_name = read_field(reader, "general.name")
|
model_name = read_field(reader, "general.name")
|
||||||
|
|
||||||
updated_architecture = None
|
updated_architecture = None
|
||||||
|
@ -298,6 +298,7 @@ class GgufModelTests(unittest.TestCase):
|
|||||||
gemma2_model_id = "bartowski/gemma-2-2b-it-GGUF"
|
gemma2_model_id = "bartowski/gemma-2-2b-it-GGUF"
|
||||||
original_gemma3_text_model_id = "google/gemma-3-1b-it"
|
original_gemma3_text_model_id = "google/gemma-3-1b-it"
|
||||||
original_gemma3_vision_model_id = "google/gemma-3-4b-it"
|
original_gemma3_vision_model_id = "google/gemma-3-4b-it"
|
||||||
|
gemma3_qat_model_id = "google/gemma-3-1b-it-qat-q4_0-gguf"
|
||||||
gemma3_text_model_id = "unsloth/gemma-3-1b-it-GGUF"
|
gemma3_text_model_id = "unsloth/gemma-3-1b-it-GGUF"
|
||||||
gemma3_vision_model_id = "unsloth/gemma-3-4b-it-GGUF"
|
gemma3_vision_model_id = "unsloth/gemma-3-4b-it-GGUF"
|
||||||
|
|
||||||
@ -329,7 +330,7 @@ class GgufModelTests(unittest.TestCase):
|
|||||||
q3_k_gemma2_model_id = "gemma-2-2b-it-Q3_K_L.gguf"
|
q3_k_gemma2_model_id = "gemma-2-2b-it-Q3_K_L.gguf"
|
||||||
q8_0_gemma2_model_id = "gemma-2-2b-it-Q8_0.gguf"
|
q8_0_gemma2_model_id = "gemma-2-2b-it-Q8_0.gguf"
|
||||||
fp32_gemma2_model_id = "gemma-2-2b-it-f32.gguf"
|
fp32_gemma2_model_id = "gemma-2-2b-it-f32.gguf"
|
||||||
q2_k_gemma3_text_model_id = "gemma-3-1b-it-Q2_K.gguf"
|
q4_0_gemma3_qat_model_id = "gemma-3-1b-it-q4_0.gguf"
|
||||||
bf16_gemma3_text_model_id = "gemma-3-1b-it-BF16.gguf"
|
bf16_gemma3_text_model_id = "gemma-3-1b-it-BF16.gguf"
|
||||||
bf16_gemma3_vision_model_id = "gemma-3-4b-it-BF16.gguf"
|
bf16_gemma3_vision_model_id = "gemma-3-4b-it-BF16.gguf"
|
||||||
|
|
||||||
@ -889,19 +890,20 @@ class GgufModelTests(unittest.TestCase):
|
|||||||
else:
|
else:
|
||||||
raise ValueError(f"Layer {layer_name} is not presented in GGUF model")
|
raise ValueError(f"Layer {layer_name} is not presented in GGUF model")
|
||||||
|
|
||||||
|
@require_read_token
|
||||||
@unittest.skipUnless(is_gguf_available("0.16.0"), "test requires gguf version >= 0.16.0")
|
@unittest.skipUnless(is_gguf_available("0.16.0"), "test requires gguf version >= 0.16.0")
|
||||||
def test_gemma3_text_q2_k(self):
|
def test_gemma3_qat_q4_0(self):
|
||||||
model = AutoModelForCausalLM.from_pretrained(
|
model = AutoModelForCausalLM.from_pretrained(
|
||||||
self.gemma3_text_model_id,
|
self.gemma3_qat_model_id,
|
||||||
gguf_file=self.q2_k_gemma3_text_model_id,
|
gguf_file=self.q4_0_gemma3_qat_model_id,
|
||||||
torch_dtype=torch.float16,
|
torch_dtype=torch.float16,
|
||||||
)
|
)
|
||||||
|
|
||||||
tokenizer = AutoTokenizer.from_pretrained(self.gemma3_text_model_id, gguf_file=self.q2_k_gemma3_text_model_id)
|
tokenizer = AutoTokenizer.from_pretrained(self.gemma3_qat_model_id, gguf_file=self.q4_0_gemma3_qat_model_id)
|
||||||
text = tokenizer(self.example_text, return_tensors="pt")["input_ids"]
|
text = tokenizer(self.example_text, return_tensors="pt")["input_ids"]
|
||||||
out = model.generate(text, max_new_tokens=10)
|
out = model.generate(text, max_new_tokens=10)
|
||||||
|
|
||||||
EXPECTED_TEXT = "Hello,\n\nI'm looking for a small,"
|
EXPECTED_TEXT = 'Hello with the prompt, "What is the best way'
|
||||||
self.assertEqual(tokenizer.decode(out[0], skip_special_tokens=True), EXPECTED_TEXT)
|
self.assertEqual(tokenizer.decode(out[0], skip_special_tokens=True), EXPECTED_TEXT)
|
||||||
|
|
||||||
@require_read_token
|
@require_read_token
|
||||||
|
Loading…
Reference in New Issue
Block a user