Pytorch学习

注意:所有文章结合读,不要只读一篇文章!!!!

一:张量的学习

(一)基础数据结构--张量

https://blog.csdn.net/oldmao_2001/article/details/102534415

https://blog.csdn.net/oldmao_2001/article/details/102546607

(二)张量学习(含代码)

https://blog.zhangxiann.com/202002052039/

https://blog.zhangxiann.com/202002082037/

(三)计算图机制

https://blog.zhangxiann.com/202002112035/

https://blog.csdn.net/oldmao_2001/article/details/102546607

(四)autograd 与逻辑回归

https://blog.zhangxiann.com/202002152033/

https://blog.csdn.net/oldmao_2001/article/details/102638970

(五)补充

1.pytorch中的.detach和.data深入详解

相同点:

tensor.data和tensor.detach() 都是变量从图中分离,但而这都是“原位操作 inplace operation”。

不同点:

1).data 是一个属性,二.detach()是一个方法;

(2).data 是不安全的,.detach()是安全的。

2.饱和和非饱和激活函数

二:数据预处理与加载

(一)数据加载DataLoader 与 DataSet

https://blog.zhangxiann.com/202002192017/自定义dataset,加载数据)

https://blog.csdn.net/oldmao_2001/article/details/102661974

(二)数据预处理与数据增强

https://blog.zhangxiann.com/202002212045/(预处理了解)

https://blog.csdn.net/oldmao_2001/article/details/102718002自定义transform方法

https://blog.zhangxiann.com/202002272047/(数据增强)

(三)补充

为什么pytorch中transforms.ToTensor要把(H,W,C)的矩阵转为(C,H,W)?

三:模型构建

(一)模型创建步骤与 nn.Module

https://blog.csdn.net/oldmao_2001/article/details/102787546

https://blog.zhangxiann.com/202003012001/

(二)nn网络层

https://blog.csdn.net/oldmao_2001/article/details/102844727(总)

https://blog.zhangxiann.com/202003032009/(卷积层)

https://blog.zhangxiann.com/202003072007/(池化层、线性层和激活函数层)

四:模型训练

(一)权值初始化(两篇都要看)---xavier与kaiming

https://blog.zhangxiann.com/202003092013/(着重:正态)

https://zhuanlan.zhihu.com/p/148034113(着重:均匀)

https://blog.csdn.net/oldmao_2001/article/details/102895144

(二)损失函数

(1)交叉熵

交叉熵=信息熵+相对熵https://www.zhihu.com/question/41252833

https://www.cnblogs.com/JeasonIsCoding/p/10171201.html

https://www.cnblogs.com/ssyfj/p/13966848.html(重点:最后都是二分类和多分类)

注意:https://blog.csdn.net/oldmao_2001/article/details/102895144中的交叉熵求解也是对的,可以推导

(2)损失函数

https://blog.csdn.net/oldmao_2001/article/details/102895144

https://blog.csdn.net/oldmao_2001/article/details/102947877

(三)优化器

https://blog.zhangxiann.com/202003172017/

https://blog.csdn.net/oldmao_2001/article/details/102947877

https://blog.csdn.net/oldmao_2001/article/details/102967125

(四)学习率调整

https://blog.csdn.net/oldmao_2001/article/details/103047336(学习率调整)

(五)补充

torch代码解析 为什么要使用optimizer.zero_grad()

五:可视化

(一)tensorboard

https://blog.zhangxiann.com/202003192045/

https://blog.csdn.net/oldmao_2001/article/details/103063776

(二)Hook函数(重点)

https://blog.zhangxiann.com/202003232051/

https://blog.csdn.net/oldmao_2001/article/details/103123751

六:正则化

(一)正则化与dropout

https://blog.zhangxiann.com/202003272049/(model.eval() 和 model.trian())

(二)Normalization

(1)原理

https://blog.csdn.net/qq_23262411/article/details/100175943(重点)

(2)如何区分并记住常见的几种 Normalization 算法(重点)

https://zhuanlan.zhihu.com/p/69659844

https://blog.zhangxiann.com/202004011919/

(3)BN1D,BN2D,BN3D

https://blog.csdn.net/Elvirangel/article/details/105770141

 (4)其他

https://blog.csdn.net/oldmao_2001/article/details/103182891

https://blog.zhangxiann.com/202004011919/

七:模型其他操作

(一)模型保存与加载

https://blog.zhangxiann.com/202004051903/(含模型断点续训练)(重点)

https://blog.zhangxiann.com/202003172017/(回顾优化器中的state_dict方法)

https://blog.csdn.net/oldmao_2001/article/details/103235986(不含代码,有图)

(二)模型迁移学习(两篇一定结合看)(含GPU使用

https://blog.zhangxiann.com/202004091911/(代码拆分解读)

https://blog.csdn.net/oldmao_2001/article/details/103235986(完整代码)

(三)使用GPU训练模型及常见错误(结合两篇读)

https://blog.zhangxiann.com/202004151915/(命令解释清楚)

https://blog.csdn.net/oldmao_2001/article/details/103269437(代码全面)

(四)补充

数学运算有in-place和none-in-place两种形式

八:残差网络 

ResNet网络结构(视频)

https://www.cnblogs.com/alanma/p/6877166.html

https://blog.csdn.net/u013181595/article/details/80990930

https://blog.zhangxiann.com/202004171947/(不错)

posted @ 2020-12-10 14:21  山上有风景  阅读(234)  评论(0编辑  收藏  举报