From 972137977096c687dcbb54df7c33f69359474d7e Mon Sep 17 00:00:00 2001 From: Giuseppe Franco Date: Wed, 15 Jan 2025 19:15:11 +0100 Subject: [PATCH] Fix (brevitas_examples/llm): remove unnecessary checkpointing (#1161) --- src/brevitas_examples/llm/main.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/src/brevitas_examples/llm/main.py b/src/brevitas_examples/llm/main.py index 21641a819..98c37027c 100644 --- a/src/brevitas_examples/llm/main.py +++ b/src/brevitas_examples/llm/main.py @@ -517,10 +517,6 @@ def quantize_llm(args): apply_bias_correction(model, calibration_loader) print("Bias correction applied.") - if args.checkpoint_name is not None: - print(f"Saving checkpoint to {args.checkpoint_name}") - torch.save(model.state_dict(), args.checkpoint_name) - if args.eval and not args.no_quantize: print("Model eval...") with torch.no_grad(), quant_inference_mode(model):