From d515d0f0746c76bd58e138ec882063c1ba768960 Mon Sep 17 00:00:00 2001 From: Manuel Schmid Date: Fri, 26 Jan 2024 02:06:17 +0100 Subject: [PATCH] fix: remove unnecessary fp32 / fp16 handling --- extras/GroundingDINO/util/inference.py | 6 ------ 1 file changed, 6 deletions(-) diff --git a/extras/GroundingDINO/util/inference.py b/extras/GroundingDINO/util/inference.py index e4b723a5..3f61e946 100644 --- a/extras/GroundingDINO/util/inference.py +++ b/extras/GroundingDINO/util/inference.py @@ -18,7 +18,6 @@ class GroundingDinoModel(Model): self.model = None self.load_device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') self.offload_device = torch.device('cpu') - self.dtype = torch.float32 def predict_with_caption( self, @@ -36,14 +35,9 @@ class GroundingDinoModel(Model): self.load_device = model_management.text_encoder_device() self.offload_device = model_management.text_encoder_offload_device() - self.dtype = torch.float32 model.to(self.offload_device) - if model_management.should_use_fp16(device=self.load_device): - model.half() - self.dtype = torch.float16 - self.model = ModelPatcher(model, load_device=self.load_device, offload_device=self.offload_device) model_management.load_model_gpu(self.model)