Skip to content

Instantly share code, notes, and snippets.

@mjdietzx
Created August 3, 2017 20:27
Show Gist options
  • Star 6 You must be signed in to star a gist
  • Fork 2 You must be signed in to fork a gist
  • Save mjdietzx/50d3c26f1fd543f1808ffffacc987cbf to your computer and use it in GitHub Desktop.
Save mjdietzx/50d3c26f1fd543f1808ffffacc987cbf to your computer and use it in GitHub Desktop.
Cross entropy loss pytorch implementation
import torch
from torch import autograd
from torch import nn
class CrossEntropyLoss(nn.Module):
"""
This criterion (`CrossEntropyLoss`) combines `LogSoftMax` and `NLLLoss` in one single class.
NOTE: Computes per-element losses for a mini-batch (instead of the average loss over the entire mini-batch).
"""
log_softmax = nn.LogSoftmax()
def __init__(self, class_weights):
super().__init__()
self.class_weights = autograd.Variable(torch.FloatTensor(class_weights).cuda())
def forward(self, logits, target):
log_probabilities = self.log_softmax(logits)
# NLLLoss(x, class) = -weights[class] * x[class]
return -self.class_weights.index_select(0, target) * log_probabilities.index_select(-1, target).diag()
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment