mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-24 23:08:57 +06:00
64 lines
2.9 KiB
Python
64 lines
2.9 KiB
Python
"""
|
|
Framework agnostic tests for generate()-related methods.
|
|
"""
|
|
|
|
import numpy as np
|
|
|
|
from transformers import AutoTokenizer
|
|
|
|
|
|
class GenerationIntegrationTestsMixin:
|
|
|
|
# To be populated by the child classes
|
|
framework_dependent_parameters = {
|
|
"AutoModelForSeq2SeqLM": None,
|
|
"LogitsProcessorList": None,
|
|
"MinLengthLogitsProcessor": None,
|
|
"create_tensor_fn": None,
|
|
"return_tensors": None,
|
|
}
|
|
|
|
def test_validate_generation_inputs(self):
|
|
model_cls = self.framework_dependent_parameters["AutoModelForSeq2SeqLM"]
|
|
return_tensors = self.framework_dependent_parameters["return_tensors"]
|
|
create_tensor_fn = self.framework_dependent_parameters["create_tensor_fn"]
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5")
|
|
model = model_cls.from_pretrained("hf-internal-testing/tiny-random-t5")
|
|
|
|
encoder_input_str = "Hello world"
|
|
input_ids = tokenizer(encoder_input_str, return_tensors=return_tensors).input_ids
|
|
|
|
# typos are quickly detected (the correct argument is `do_sample`)
|
|
with self.assertRaisesRegex(ValueError, "do_samples"):
|
|
model.generate(input_ids, do_samples=True)
|
|
|
|
# arbitrary arguments that will not be used anywhere are also not accepted
|
|
with self.assertRaisesRegex(ValueError, "foo"):
|
|
fake_model_kwargs = {"foo": "bar"}
|
|
model.generate(input_ids, **fake_model_kwargs)
|
|
|
|
# however, valid model_kwargs are accepted
|
|
valid_model_kwargs = {"attention_mask": create_tensor_fn(np.zeros_like(input_ids))}
|
|
model.generate(input_ids, **valid_model_kwargs)
|
|
|
|
def test_custom_logits_processor(self):
|
|
model_cls = self.framework_dependent_parameters["AutoModelForSeq2SeqLM"]
|
|
logits_processor_list_cls = self.framework_dependent_parameters["LogitsProcessorList"]
|
|
min_length_logits_processor_cls = self.framework_dependent_parameters["MinLengthLogitsProcessor"]
|
|
return_tensors = self.framework_dependent_parameters["return_tensors"]
|
|
|
|
bart_tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-bart")
|
|
article = """Justin Timberlake and Jessica Biel, welcome to parenthood."""
|
|
bart_model = model_cls.from_pretrained("hf-internal-testing/tiny-random-bart", min_length=1)
|
|
input_ids = bart_tokenizer(article, return_tensors=return_tensors).input_ids
|
|
|
|
logits_processor = logits_processor_list_cls()
|
|
logits_processor.append(min_length_logits_processor_cls(min_length=10, eos_token_id=0))
|
|
# it should not be allowed to both define `min_length` via config and `logits_processor` list
|
|
with self.assertRaises(ValueError):
|
|
bart_model.generate(input_ids, logits_processor=logits_processor)
|
|
|
|
bart_model.config.min_length = None
|
|
bart_model.generate(input_ids, logits_processor=logits_processor)
|