[docs]deffuse_conv_bn_eval(conv:ConvT,bn:torch.nn.modules.batchnorm._BatchNorm,transpose:bool=False)->ConvT:r"""Fuse a convolutional module and a BatchNorm module into a single, new convolutional module. Args: conv (torch.nn.modules.conv._ConvNd): A convolutional module. bn (torch.nn.modules.batchnorm._BatchNorm): A BatchNorm module. transpose (bool, optional): If True, transpose the convolutional weight. Defaults to False. Returns: torch.nn.modules.conv._ConvNd: The fused convolutional module. .. note:: Both ``conv`` and ``bn`` must be in eval mode, and ``bn`` must have its running buffers computed. """assertnot(conv.trainingorbn.training),"Fusion only for eval!"fused_conv=copy.deepcopy(conv)assertbn.running_meanisnotNoneandbn.running_varisnotNonefused_conv.weight,fused_conv.bias=fuse_conv_bn_weights(fused_conv.weight,fused_conv.bias,bn.running_mean,bn.running_var,bn.eps,bn.weight,bn.bias,transpose)returnfused_conv
[docs]deffuse_conv_bn_weights(conv_w:torch.Tensor,conv_b:Optional[torch.Tensor],bn_rm:torch.Tensor,bn_rv:torch.Tensor,bn_eps:float,bn_w:Optional[torch.Tensor],bn_b:Optional[torch.Tensor],transpose:bool=False)->Tuple[torch.nn.Parameter,torch.nn.Parameter]:r"""Fuse convolutional module parameters and BatchNorm module parameters into new convolutional module parameters. Args: conv_w (torch.Tensor): Convolutional weight. conv_b (Optional[torch.Tensor]): Convolutional bias. bn_rm (torch.Tensor): BatchNorm running mean. bn_rv (torch.Tensor): BatchNorm running variance. bn_eps (float): BatchNorm epsilon. bn_w (Optional[torch.Tensor]): BatchNorm weight. bn_b (Optional[torch.Tensor]): BatchNorm bias. transpose (bool, optional): If True, transpose the conv weight. Defaults to False. Returns: Tuple[torch.nn.Parameter, torch.nn.Parameter]: Fused convolutional weight and bias. """conv_weight_dtype=conv_w.dtypeconv_bias_dtype=conv_b.dtypeifconv_bisnotNoneelseconv_weight_dtypeifconv_bisNone:conv_b=torch.zeros_like(bn_rm)ifbn_wisNone:bn_w=torch.ones_like(bn_rm)ifbn_bisNone:bn_b=torch.zeros_like(bn_rm)bn_var_rsqrt=torch.rsqrt(bn_rv+bn_eps)iftranspose:shape=[1,-1]+[1]*(len(conv_w.shape)-2)else:shape=[-1,1]+[1]*(len(conv_w.shape)-2)fused_conv_w=(conv_w*(bn_w*bn_var_rsqrt).reshape(shape)).to(dtype=conv_weight_dtype)fused_conv_b=((conv_b-bn_rm)*bn_var_rsqrt*bn_w+bn_b).to(dtype=conv_bias_dtype)return(torch.nn.Parameter(fused_conv_w,conv_w.requires_grad),torch.nn.Parameter(fused_conv_b,conv_b.requires_grad))
[docs]deffuse_linear_bn_eval(linear:LinearT,bn:torch.nn.modules.batchnorm._BatchNorm)->LinearT:r"""Fuse a linear module and a BatchNorm module into a single, new linear module. Args: linear (torch.nn.Linear): A Linear module. bn (torch.nn.modules.batchnorm._BatchNorm): A BatchNorm module. Returns: torch.nn.Linear: The fused linear module. .. note:: Both ``linear`` and ``bn`` must be in eval mode, and ``bn`` must have its running buffers computed. """assertnot(linear.trainingorbn.training),"Fusion only for eval!"fused_linear=copy.deepcopy(linear)assertbn.running_meanisnotNoneandbn.running_varisnotNonefused_linear.weight,fused_linear.bias=fuse_linear_bn_weights(fused_linear.weight,fused_linear.bias,bn.running_mean,bn.running_var,bn.eps,bn.weight,bn.bias)returnfused_linear
[docs]deffuse_linear_bn_weights(linear_w:torch.Tensor,linear_b:Optional[torch.Tensor],bn_rm:torch.Tensor,bn_rv:torch.Tensor,bn_eps:float,bn_w:torch.Tensor,bn_b:torch.Tensor,)->Tuple[torch.nn.Parameter,torch.nn.Parameter]:r"""Fuse linear module parameters and BatchNorm module parameters into new linear module parameters. Args: linear_w (torch.Tensor): Linear weight. linear_b (Optional[torch.Tensor]): Linear bias. bn_rm (torch.Tensor): BatchNorm running mean. bn_rv (torch.Tensor): BatchNorm running variance. bn_eps (float): BatchNorm epsilon. bn_w (torch.Tensor): BatchNorm weight. bn_b (torch.Tensor): BatchNorm bias. transpose (bool, optional): If True, transpose the conv weight. Defaults to False. Returns: Tuple[torch.nn.Parameter, torch.nn.Parameter]: Fused linear weight and bias. """iflinear_bisNone:linear_b=torch.zeros_like(bn_rm)bn_scale=bn_w*torch.rsqrt(bn_rv+bn_eps)fused_w=linear_w*bn_scale.unsqueeze(-1)fused_b=(linear_b-bn_rm)*bn_scale+bn_breturntorch.nn.Parameter(fused_w,linear_w.requires_grad),torch.nn.Parameter(fused_b,linear_b.requires_grad)
Docs
Access comprehensive developer documentation for PyTorch
To analyze traffic and optimize your experience, we serve cookies on this site. By clicking or navigating, you agree to allow our usage of cookies. As the current maintainers of this site, Facebook’s Cookies Policy applies. Learn more, including about available controls: Cookies Policy.