mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-03 12:50:06 +06:00
Support loading Gemma3 QAT GGUF models (#37649)
* fix gemma3 qat gguf support Signed-off-by: isotr0py <2037008807@qq.com> * update test Signed-off-by: isotr0py <2037008807@qq.com> * make ruff happy Signed-off-by: isotr0py <2037008807@qq.com> --------- Signed-off-by: isotr0py <2037008807@qq.com> Co-authored-by: Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
This commit is contained in:
parent
7eb1107cc2
commit
c69e23455d
@ -258,6 +258,8 @@ TENSOR_PROCESSORS = {
|
||||
|
||||
|
||||
def read_field(reader, field):
|
||||
if field not in reader.fields:
|
||||
return []
|
||||
value = reader.fields[field]
|
||||
return [_gguf_parse_value(value.parts[_data_index], value.types) for _data_index in value.data]
|
||||
|
||||
@ -369,6 +371,7 @@ def load_gguf_checkpoint(gguf_checkpoint_path, return_tensors=False, model_to_lo
|
||||
parsed_parameters = {k: {} for k in GGUF_TO_TRANSFORMERS_MAPPING}
|
||||
|
||||
architecture = read_field(reader, "general.architecture")[0]
|
||||
# NOTE: Some GGUF checkpoints may miss `general.name` field in metadata
|
||||
model_name = read_field(reader, "general.name")
|
||||
|
||||
updated_architecture = None
|
||||
|
@ -298,6 +298,7 @@ class GgufModelTests(unittest.TestCase):
|
||||
gemma2_model_id = "bartowski/gemma-2-2b-it-GGUF"
|
||||
original_gemma3_text_model_id = "google/gemma-3-1b-it"
|
||||
original_gemma3_vision_model_id = "google/gemma-3-4b-it"
|
||||
gemma3_qat_model_id = "google/gemma-3-1b-it-qat-q4_0-gguf"
|
||||
gemma3_text_model_id = "unsloth/gemma-3-1b-it-GGUF"
|
||||
gemma3_vision_model_id = "unsloth/gemma-3-4b-it-GGUF"
|
||||
|
||||
@ -329,7 +330,7 @@ class GgufModelTests(unittest.TestCase):
|
||||
q3_k_gemma2_model_id = "gemma-2-2b-it-Q3_K_L.gguf"
|
||||
q8_0_gemma2_model_id = "gemma-2-2b-it-Q8_0.gguf"
|
||||
fp32_gemma2_model_id = "gemma-2-2b-it-f32.gguf"
|
||||
q2_k_gemma3_text_model_id = "gemma-3-1b-it-Q2_K.gguf"
|
||||
q4_0_gemma3_qat_model_id = "gemma-3-1b-it-q4_0.gguf"
|
||||
bf16_gemma3_text_model_id = "gemma-3-1b-it-BF16.gguf"
|
||||
bf16_gemma3_vision_model_id = "gemma-3-4b-it-BF16.gguf"
|
||||
|
||||
@ -889,19 +890,20 @@ class GgufModelTests(unittest.TestCase):
|
||||
else:
|
||||
raise ValueError(f"Layer {layer_name} is not presented in GGUF model")
|
||||
|
||||
@require_read_token
|
||||
@unittest.skipUnless(is_gguf_available("0.16.0"), "test requires gguf version >= 0.16.0")
|
||||
def test_gemma3_text_q2_k(self):
|
||||
def test_gemma3_qat_q4_0(self):
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
self.gemma3_text_model_id,
|
||||
gguf_file=self.q2_k_gemma3_text_model_id,
|
||||
self.gemma3_qat_model_id,
|
||||
gguf_file=self.q4_0_gemma3_qat_model_id,
|
||||
torch_dtype=torch.float16,
|
||||
)
|
||||
|
||||
tokenizer = AutoTokenizer.from_pretrained(self.gemma3_text_model_id, gguf_file=self.q2_k_gemma3_text_model_id)
|
||||
tokenizer = AutoTokenizer.from_pretrained(self.gemma3_qat_model_id, gguf_file=self.q4_0_gemma3_qat_model_id)
|
||||
text = tokenizer(self.example_text, return_tensors="pt")["input_ids"]
|
||||
out = model.generate(text, max_new_tokens=10)
|
||||
|
||||
EXPECTED_TEXT = "Hello,\n\nI'm looking for a small,"
|
||||
EXPECTED_TEXT = 'Hello with the prompt, "What is the best way'
|
||||
self.assertEqual(tokenizer.decode(out[0], skip_special_tokens=True), EXPECTED_TEXT)
|
||||
|
||||
@require_read_token
|
||||
|
Loading…
Reference in New Issue
Block a user