提交 6da60745 编写于 作者: V Varuna Jayasiri

loop timing

上级 ebbe704d
......@@ -493,7 +493,7 @@ class Configs(BaseConfigs):
# Loop through epochs
for epoch in monit.loop(self.epochs):
# Loop through the dataset
for i, batch in enumerate(self.dataloader):
for i, batch in monit.enum('Train', self.dataloader):
# Move images to the device
data_x, data_y = batch['x'].to(self.device), batch['y'].to(self.device)
......@@ -528,6 +528,8 @@ class Configs(BaseConfigs):
# Update learning rates
self.generator_lr_scheduler.step()
self.discriminator_lr_scheduler.step()
# New line
tracker.new_line()
def optimize_generators(self, data_x: torch.Tensor, data_y: torch.Tensor, true_labels: torch.Tensor):
"""
......
......@@ -5,7 +5,7 @@ with open("readme.md", "r") as f:
setuptools.setup(
name='labml-nn',
version='0.4.80',
version='0.4.81',
author="Varuna Jayasiri, Nipun Wijerathne",
author_email="vpjayasiri@gmail.com, hnipun@gmail.com",
description="A collection of PyTorch implementations of neural network architectures and layers.",
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册