摘要: 转自:https://discuss.pytorch.org/t/why-do-we-need-to-do-loss-cuda-when-we-we-have-already-done-model-cuda/91023/5 https://discuss.pytorch.org/t/move-the 阅读全文
posted @ 2021-11-11 22:28 lypbendlf 阅读(743) 评论(0) 推荐(0) 编辑
摘要: 转自:https://www.jb51.net/article/213149.htm 1.多个loss x = torch.tensor(2.0, requires_grad=True) y = x**2 z = x # 反向传播 y.backward() x.grad tensor(4.) z.b 阅读全文
posted @ 2021-11-11 22:20 lypbendlf 阅读(467) 评论(0) 推荐(0) 编辑
摘要: 1.训练报错 使用BCE损失时,出现的问题包括: 报错 参数batch_size | epoch | hidden_size | lr_D | lr_DZ | lr_Eref | lr_model | z_dim 'ViewBackward' returned nan values 8 | 50 | 阅读全文
posted @ 2021-11-11 20:01 lypbendlf 阅读(1361) 评论(0) 推荐(0) 编辑