![Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium](https://miro.medium.com/v2/resize:fit:469/1*8Kvne7teaEVoq5X78DyRMA.png)
Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium
![Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names](https://gombru.github.io/assets/cross_entropy_loss/intro.png)
Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names
![neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow](https://i.stack.imgur.com/e6gKc.png)
neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow
![Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums](https://discuss.pytorch.org/uploads/default/original/2X/2/2ad4119a40ee6e24f006aabae0f6d0981a20a9cf.png)
Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums
GitHub - mochangheng/crnn-crossentropy-pytorch: Cross-entropy CRNN implementation in PyTorch for state-of-the-art short-length OCR
![Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums](https://discuss.pytorch.org/uploads/default/original/2X/4/4ac4609aee5df6b686796f7490dcd242f463fe5b.png)
Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums
![machine learning - Cross Entropy in PyTorch is different from what I learnt (Not about logit input, but about the loss for every node) - Cross Validated machine learning - Cross Entropy in PyTorch is different from what I learnt (Not about logit input, but about the loss for every node) - Cross Validated](https://i.stack.imgur.com/zua3x.png)