From 38deac1d40dc01e9e0c217be7621003a88a13bfd Mon Sep 17 00:00:00 2001 From: xinhe Date: Tue, 5 Nov 2024 15:38:18 +0800 Subject: [PATCH] Revert "Fix bug when loading 4bit checkpoint quantized in INC (#1447)" This reverts commit 4bdf4348702c99e1efeeb73c8ef2c23b924c0097. --- examples/text-generation/utils.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/examples/text-generation/utils.py b/examples/text-generation/utils.py index 698a67ec5a..cb734071b0 100644 --- a/examples/text-generation/utils.py +++ b/examples/text-generation/utils.py @@ -269,9 +269,6 @@ def setup_model(args, model_dtype, model_kwargs, logger): original_model=org_model, **model_kwargs, ) - # TODO: This will be removed in v1.19 Synapse release - # the loaded model should have the same dtype as original_model - model = model.to(model_kwargs["torch_dtype"]) else: if args.assistant_model is not None: assistant_model = AutoModelForCausalLM.from_pretrained(