From c69e342ee4f573550b148444b9ff7db997505075 Mon Sep 17 00:00:00 2001 From: Sean Sube Date: Sun, 19 Feb 2023 07:53:20 -0600 Subject: [PATCH] apply sonar lint --- api/onnx_web/chain/blend_inpaint.py | 2 +- api/onnx_web/chain/source_noise.py | 2 +- api/onnx_web/convert/__main__.py | 2 +- api/onnx_web/convert/diffusion_original.py | 2 -- api/onnx_web/convert/utils.py | 4 ++-- api/onnx_web/diffusion/run.py | 2 -- api/onnx_web/server/device_pool.py | 1 - api/onnx_web/server/model_cache.py | 9 ++++----- 8 files changed, 9 insertions(+), 15 deletions(-) diff --git a/api/onnx_web/chain/blend_inpaint.py b/api/onnx_web/chain/blend_inpaint.py index bd51eb3e..a19cb8b7 100644 --- a/api/onnx_web/chain/blend_inpaint.py +++ b/api/onnx_web/chain/blend_inpaint.py @@ -113,5 +113,5 @@ def blend_inpaint( output = process_tile_order(stage.tile_order, source, SizeChart.auto, 1, [outpaint]) - logger.info("final output image size", output.size) + logger.info("final output image size: %s", output.size) return output diff --git a/api/onnx_web/chain/source_noise.py b/api/onnx_web/chain/source_noise.py index 76cc4f73..f6267b26 100644 --- a/api/onnx_web/chain/source_noise.py +++ b/api/onnx_web/chain/source_noise.py @@ -13,7 +13,7 @@ def source_noise( _job: JobContext, _server: ServerContext, _stage: StageParams, - params: ImageParams, + _params: ImageParams, source: Image.Image, *, size: Size, diff --git a/api/onnx_web/convert/__main__.py b/api/onnx_web/convert/__main__.py index 08c83caa..774a9e72 100644 --- a/api/onnx_web/convert/__main__.py +++ b/api/onnx_web/convert/__main__.py @@ -149,7 +149,7 @@ def fetch_model( if model_format is None: url = urlparse(source) ext = path.basename(url.path) - file, ext = path.splitext(ext) + _filename, ext = path.splitext(ext) if ext is not None: cache_name += ext else: diff --git a/api/onnx_web/convert/diffusion_original.py b/api/onnx_web/convert/diffusion_original.py index 4aed40b9..da00c45b 100644 --- a/api/onnx_web/convert/diffusion_original.py +++ b/api/onnx_web/convert/diffusion_original.py @@ -1397,8 +1397,6 @@ def extract_checkpoint( logger.info(result_status) - return - def convert_diffusion_original( ctx: ConversionContext, diff --git a/api/onnx_web/convert/utils.py b/api/onnx_web/convert/utils.py index 24850252..9d0f96ab 100644 --- a/api/onnx_web/convert/utils.py +++ b/api/onnx_web/convert/utils.py @@ -214,12 +214,12 @@ def load_tensor(name: str, map_location=None): except Exception as e: try: logger.warning( - "failed to load as safetensors file, falling back to torch", e + "failed to load as safetensors file, falling back to torch: %s", e ) checkpoint = torch.jit.load(name) except Exception as e: logger.warning( - "failed to load with Torch JIT, falling back to PyTorch", e + "failed to load with Torch JIT, falling back to PyTorch: %s", e ) checkpoint = torch.load(name, map_location=map_location) checkpoint = ( diff --git a/api/onnx_web/diffusion/run.py b/api/onnx_web/diffusion/run.py index 9ac97942..ebd56cb6 100644 --- a/api/onnx_web/diffusion/run.py +++ b/api/onnx_web/diffusion/run.py @@ -173,7 +173,6 @@ def run_inpaint_pipeline( fill_color: str, tile_order: str, ) -> None: - # device = job.get_device() progress = job.get_progress_callback() stage = StageParams(tile_order=tile_order) @@ -218,7 +217,6 @@ def run_upscale_pipeline( upscale: UpscaleParams, source: Image.Image, ) -> None: - # device = job.get_device() progress = job.get_progress_callback() stage = StageParams() diff --git a/api/onnx_web/server/device_pool.py b/api/onnx_web/server/device_pool.py index b935d549..ef091d8e 100644 --- a/api/onnx_web/server/device_pool.py +++ b/api/onnx_web/server/device_pool.py @@ -208,7 +208,6 @@ class DevicePoolExecutor: except ValueError as e: logger.warning("error removing pruned job from pending: %s", e) - # self.jobs[:] = [job for job in self.jobs if not job.future.done()] recent_count = len(self.recent) if recent_count > self.recent_limit: logger.debug( diff --git a/api/onnx_web/server/model_cache.py b/api/onnx_web/server/model_cache.py index b940fcf3..7b1f4728 100644 --- a/api/onnx_web/server/model_cache.py +++ b/api/onnx_web/server/model_cache.py @@ -31,11 +31,10 @@ class ModelCache: for i in range(len(self.cache)): t, k, v = self.cache[i] - if tag == t: - if key != k: - logger.debug("updating model cache: %s", tag) - self.cache[i] = (tag, key, value) - return + if tag == t and key != k: + logger.debug("updating model cache: %s", tag) + self.cache[i] = (tag, key, value) + return logger.debug("adding new model to cache: %s", tag) self.cache.append((tag, key, value))