fix(api): assume inversion tokens are embeddings for now
This commit is contained in:
parent
17a7cdae56
commit
e5862d178c
|
@ -217,8 +217,8 @@ def load_pipeline(
|
|||
|
||||
text_encoder = None
|
||||
if inversions is not None and len(inversions) > 0:
|
||||
logger.debug("blending Textual Inversions from %s", inversions)
|
||||
inversion_names, inversion_weights = zip(*inversions)
|
||||
logger.debug("blending Textual Inversions from %s", inversion_names)
|
||||
|
||||
inversion_models = [
|
||||
path.join(server.model_path, "inversion", f"{name}.ckpt")
|
||||
|
@ -233,7 +233,14 @@ def load_pipeline(
|
|||
server,
|
||||
text_encoder,
|
||||
tokenizer,
|
||||
list(zip(inversion_models, inversion_weights, inversion_names)),
|
||||
list(
|
||||
zip(
|
||||
inversion_models,
|
||||
inversion_weights,
|
||||
inversion_names,
|
||||
[None] * len(inversion_models),
|
||||
)
|
||||
),
|
||||
)
|
||||
|
||||
components["tokenizer"] = tokenizer
|
||||
|
|
|
@ -202,7 +202,9 @@ class DevicePoolExecutor:
|
|||
|
||||
for job in self.pending_jobs:
|
||||
if job.name == key:
|
||||
self.pending_jobs[:] = [job for job in self.pending_jobs if job.name != key]
|
||||
self.pending_jobs[:] = [
|
||||
job for job in self.pending_jobs if job.name != key
|
||||
]
|
||||
logger.info("cancelled pending job: %s", key)
|
||||
return True
|
||||
|
||||
|
@ -387,7 +389,8 @@ class DevicePoolExecutor:
|
|||
False,
|
||||
False,
|
||||
False,
|
||||
) for job in self.pending_jobs
|
||||
)
|
||||
for job in self.pending_jobs
|
||||
]
|
||||
)
|
||||
history.extend(
|
||||
|
@ -420,7 +423,9 @@ class DevicePoolExecutor:
|
|||
"progress update for job: %s to %s", progress.job, progress.progress
|
||||
)
|
||||
self.running_jobs[progress.job] = progress
|
||||
self.pending_jobs[:] = [job for job in self.pending_jobs if job.name != progress.job]
|
||||
self.pending_jobs[:] = [
|
||||
job for job in self.pending_jobs if job.name != progress.job
|
||||
]
|
||||
|
||||
if progress.job in self.cancelled_jobs:
|
||||
logger.debug(
|
||||
|
|
Loading…
Reference in New Issue