提问 发文

pytorch 正向与反向传播的过程 获取模型的梯度(gradient),并绘制梯度的直方图

微微菌

| 2024-03-12 15:18 203 0 0

引入所需要的库

import os
import torch
import torch.nn as nn
import torch.nn.functional as F
from collections import OrderedDict

一个简单的函数
z = 3 y 2 = 3 ( x + 2 ) 2 z = 3y^2 = 3(x+2)^2z=3y
2
=3(x+2)
2


out = mean ( z ) \text{out} = \text{mean}(z)out=mean(z)

∂ z ∂ x = 6 ( x + 2 ) 2 \frac{\partial z}{\partial x} = \frac{6(x+2)}{2}
∂x
∂z

=
2
6(x+2)


大家想想,这里的公式为什么要除2?

代码如下:

x = torch.tensor([[1., 2.]], requires_grad=True)
y = x + 2
z = 3 * y.pow(2)
out = z.mean() # you can try sum() to see what is the result.
out.backward()

print(f"x: {x}")
print(f"y->x: {y}")
print(f"z->y->x: {z}")
print(f"out: {out}")
print(f"out->z->y->x: {x.grad}")

输出如下

x: tensor([[1., 2.]], requires_grad=True)
y->x: tensor([[3., 4.]], grad_fn=<AddBackward0>)
z->y->x: tensor([[27., 48.]], grad_fn=<MulBackward0>)
out: 37.5
out->z->y->x: tensor([[ 9., 12.]])

这里解释一下,x 是定义一个tensor,并设置requires_grad=True,这个意思就是x需要计算梯度。其它的注释已经标注挺清楚的了

class ToyModel(nn.Module):
def __init__(self, in_channels, out_channels, num_classes=2):
super().__init__()
# tmp only for testing, not valid
self.tmp = nn.Conv2d(in_channels, in_channels * 2, (3, 3))
self.dim = out_channels
self.conv1 = nn.Conv2d(in_channels=in_channels,
out_channels=in_channels * 2,
kernel_size=(3, 3),
stride=(1, 1),
padding=0)
self.conv2 = nn.Conv2d(in_channels=in_channels * 2,
out_channels=out_channels,
kernel_size=(3, 3),
stride=(1, 1),
padding=0)
self.pool = nn.AdaptiveAvgPool2d(output_size=(1))
self.fc = nn.Linear(out_channels, num_classes, bias=False)

def forward(self, x):
x = F.relu(self.conv1(x))
x = F.relu(self.conv2(x))
x = self.pool(x)
x = self.fc(x.view(-1, self.dim))
return x


def get_model_histogram(model):
"""
Description:
- get norm gradients from model, and store in a OrderDict

Args:
- model: (torch.nn.Module), torch model

Returns:
- grads in OrderDict
"""

grads = OrderedDict()
for name, params in model.named_parameters():
grad = params.grad
if grad is not None:
tmp = {}
params_np = grad.numpy()
histogram, bins = np.histogram(params_np.flatten())
tmp['histogram'] = list(histogram)
tmp['bins'] = list(bins)
grads[name] = tmp
return grads


def get_model_norm_gradient(model):
"""
Description:
- get norm gradients from model, and store in a OrderDict

Args:
- model: (torch.nn.Module), torch model

Returns:
- grads in OrderDict
"""
grads = OrderedDict()
for name, params in model.named_parameters():
grad = params.grad
if grad is not None:
grads[name] = grad.norm().item()
return grads


get model norm gradient的输出如下

================================================================================
OrderedDict([('conv1.weight', 0.1473149210214615), ('conv1.bias', 0.16713829338550568), ('conv2.weight', 0.9203198552131653), ('conv2.bias', 0.5442095994949341), ('fc.weight', 1.7258217334747314)])
================================================================================
OrderedDict([('conv1.weight', 0.0349930003285408), ('conv1.bias', 0.03801438584923744), ('conv2.weight', 0.20729205012321472), ('conv2.bias', 0.12616902589797974), ('fc.weight', 0.39913201332092285)])
================================================================================
OrderedDict([('conv1.weight', 0.07812522351741791), ('conv1.bias', 0.08833323419094086), ('conv2.weight', 0.49012720584869385), ('conv2.bias', 0.2875416576862335), ('fc.weight', 0.9168939590454102)])
================================================================================
OrderedDict([('conv1.weight', 0.14530049264431), ('conv1.bias', 0.16511967778205872), ('conv2.weight', 0.9190732836723328), ('conv2.bias', 0.5398930907249451), ('fc.weight', 1.7265493869781494)])

torch.manual_seed(0)
num_data = 40
toy_model = ToyModel(3, 64, 2)
data = torch.randn(num_data, 3, 224, 224)
label = torch.randint(0, 2, (num_data,))
criterion = nn.CrossEntropyLoss()
optimizer = torch.optim.SGD(toy_model.parameters(), lr=1e-3)
toy_model.train()
for i, data in enumerate(data):
data = data.unsqueeze(0)
out = toy_model(data)
target = label[i].unsqueeze(0)
loss = criterion(out, target)
loss.backward()
if (i + 1) % 10 == 0:
print('=' * 80)
# fix 2022-04-27 histo not defined
# print(str(get_model_histogram(toy_model)))
histo = (get_model_histogram(toy_model))
print(histo)
optimizer.step()
optimizer.zero_grad()


get model histogram 输出如下,输出太多,只显示最后一条输入了

================================================================================
OrderedDict([('conv1.weight', {'histogram': [4, 2, 13, 27, 76, 22, 11, 5, 1, 1], 'bins': [-0.036256444, -0.028072663, -0.019888882, -0.0117051015, -0.0035213209, 0.0046624597, 0.012846241, 0.021030022, 0.029213801, 0.037397582, 0.045581363]}), ('conv1.bias', {'histogram': [1, 2, 0, 0, 1, 0, 1, 0, 0, 1], 'bins': [-0.028756114, -0.012518765, 0.0037185834, 0.019955931, 0.03619328, 0.05243063, 0.06866798, 0.08490533, 0.101142675, 0.11738002, 0.13361737]}), ('conv2.weight', {'histogram': [15, 10, 35, 245, 1828, 970, 230, 68, 40, 15], 'bins': [-0.07653718, -0.060686104, -0.044835035, -0.028983962, -0.013132891, 0.0027181804, 0.018569252, 0.034420323, 0.050271396, 0.066122465, 0.08197354]}), ('conv2.bias', {'histogram': [1, 0, 1, 8, 12, 28, 5, 6, 0, 3], 'bins': [-0.21087514, -0.16971013, -0.1285451, -0.0873801, -0.04621508, -0.005050063, 0.036114953, 0.07727997, 0.11844498, 0.15961, 0.20077501]}), ('fc.weight', {'histogram': [1, 7, 11, 12, 33, 33, 12, 11, 7, 1], 'bins': [-0.41966814, -0.33573452, -0.2518009, -0.16786726, -0.08393363, 0.0, 0.08393363, 0.16786726, 0.2518009, 0.33573452, 0.41966814]})])


可视化一下梯度的histogram

import matplotlib.pyplot as plt
import matplotlib

%matplotlib inline


  • 可视化conv2.weight
  • data = histo['conv2.weight']
    bins = data['bins']
    histogram = data['histogram']
    max_idx = np.argmax(histogram)
    min_idx = np.argmin(histogram)
    width = abs(bins[max_idx] - bins[min_idx])

    plt.figure(figsize=(9, 6))
    plt.bar(bins[:-1], histogram, width=width)
    plt.show()

  • 在这里插入图片描述

    • 可视化conv2.bias
    • data = histo['conv2.bias']
      bins = data['bins']
      histogram = data['histogram']
      max_idx = np.argmax(histogram)
      min_idx = np.argmin(histogram)
      width = abs(bins[max_idx] - bins[min_idx])

      plt.figure(figsize=(9, 6))
      plt.bar(bins[:-1], histogram, width=width)
      plt.show()

    • 在这里插入图片描述

  • 本文为二次转载,如侵权请联系删除。



收藏 0
分享
分享方式
微信

评论

游客

全部 0条评论

轻松设计高效搭建,减少3倍设计改稿与开发运维工作量

开始免费试用 预约演示

扫一扫关注公众号 扫一扫联系客服

©Copyrights 2016-2022 杭州易知微科技有限公司 浙ICP备2021017017号-3 浙公网安备33011002011932号

互联网信息服务业务 合字B2-20220090

400-8505-905 复制
免费试用
微信社区
易知微-数据可视化
微信扫一扫入群