mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 05:10:06 +06:00
parent
101b3fa4ea
commit
9a962dd9ed
@ -11,7 +11,7 @@
|
|||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
import gc
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, GenerationConfig, QuarkConfig
|
from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, GenerationConfig, QuarkConfig
|
||||||
@ -77,6 +77,14 @@ class QuarkTest(unittest.TestCase):
|
|||||||
device_map=cls.device_map,
|
device_map=cls.device_map,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def tearDown(self):
|
||||||
|
r"""
|
||||||
|
TearDown function needs to be called at the end of each test to free the GPU memory and cache, also to
|
||||||
|
avoid unexpected behaviors. Please see: https://discuss.pytorch.org/t/how-can-we-release-gpu-memory-cache/14530/27
|
||||||
|
"""
|
||||||
|
gc.collect()
|
||||||
|
torch.cuda.empty_cache()
|
||||||
|
|
||||||
def test_memory_footprint(self):
|
def test_memory_footprint(self):
|
||||||
mem_quantized = self.quantized_model.get_memory_footprint()
|
mem_quantized = self.quantized_model.get_memory_footprint()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user