提交 522a8b9f 编写于 作者: A Aarni Koskela

Add a status logger in modules.shared

上级 e4303443
......@@ -4,6 +4,7 @@ import os
import sys
import threading
import time
import logging
import gradio as gr
import torch
......@@ -18,6 +19,8 @@ from modules.paths_internal import models_path, script_path, data_path, sd_confi
from ldm.models.diffusion.ddpm import LatentDiffusion
from typing import Optional
log = logging.getLogger(__name__)
demo = None
parser = cmd_args.parser
......@@ -144,12 +147,15 @@ class State:
def request_restart(self) -> None:
self.interrupt()
self.server_command = "restart"
log.info("Received restart request")
def skip(self):
self.skipped = True
log.info("Received skip request")
def interrupt(self):
self.interrupted = True
log.info("Received interrupt request")
def nextjob(self):
if opts.live_previews_enable and opts.show_progress_every_n_steps == -1:
......@@ -189,8 +195,11 @@ class State:
self.time_start = time.time()
self.job = job
devices.torch_gc()
log.info("Starting job %s", job)
def end(self):
duration = time.time() - self.time_start
log.info("Ending job %s (%.2f seconds)", self.job, duration)
self.job = ""
self.job_count = 0
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册