Files
clip-symnets/models/DomainClassifierSource.py
2024-05-21 19:41:56 +08:00

49 lines
1.7 KiB
Python

import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.autograd import Variable
def _assert_no_grad(variable):
assert not variable.requires_grad, \
"nn criterions don't compute the gradient w.r.t. targets - please " \
"mark these variables as volatile or not requiring gradients"
class _Loss(nn.Module):
def __init__(self, size_average=True):
super(_Loss, self).__init__()
self.size_average = size_average
class _WeightedLoss(_Loss):
def __init__(self, weight=None, size_average=True):
super(_WeightedLoss, self).__init__(size_average)
self.register_buffer('weight', weight)
class DClassifierForSource(_WeightedLoss):
def __init__(self, weight=None, size_average=True, ignore_index=-100, reduce=True, nClass=10):
super(DClassifierForSource, self).__init__(weight, size_average)
self.nClass = nClass
def forward(self, input):
# _assert_no_grad(target)
batch_size = input.size(0)
prob = F.softmax(input, dim=1)
# prob = F.sigmoid(input)
c = prob.data[:, :self.nClass].sum(1)
d = (prob.data[:, :self.nClass].sum(1) == 0).sum()
if (prob.data[:, :self.nClass].sum(1) == 0).sum() != 0: ########### in case of log(0)
soft_weight = torch.FloatTensor(batch_size).fill_(0)
soft_weight[prob[:, :self.nClass].sum(1).data.cpu() == 0] = 1e-6
soft_weight_var = Variable(soft_weight).cuda()
loss = -((prob[:, :self.nClass].sum(1) + soft_weight_var).log().mean())
else:
a = prob[:, :self.nClass].sum(1) # 求数组每一行的和
b = a.log()
c = b.mean()
loss = -(prob[:, :self.nClass].sum(1).log().mean())
return loss