Cross-Entropy Loss Calculator
Use this calculator to evaluate the cross-entropy loss between your true labels (one-hot) and model predicted probabilities for classification problems.

What this calculator is doing
This calculator computes the cross-entropy loss, a standard loss function in classification tasks, using:
\[
\text{Loss} = - \sum_{i=1}^{n} y_i \cdot \log_b(p_i)
\]
Where:
- \( y_i \): True label (1 or 0, from one-hot vector)
- \( p_i \): Predicted probability for class \( i \)
- \( b \): Logarithm base (e = natural log, 2 = bits, 10 = common log)
Only the non-zero \( y_i \) values are included, meaning typically only one term in the sum.
This metric is used to measure the dissimilarity between the true distribution (label) and the predicted
distribution, where lower values indicate better predictions.
Choose from base e, 2, or 10 depending on your application (e.g., entropy in nats, bits, or bans).
Looking for similar calculators?
Check out the following calculators as well! They provide similar solutions to this one.
Want to reuse this calculator?
You can embed this calculator to your own website using the following endpoint: https://www.nichecalcs.com/embed/cross_entropy_loss
Find out more about our Services
Disclaimer: These calculators are provided for informational purposes only. Always verify your designs against relevant engineering standards and consult a qualified professional. We do not take responsibility for any errors or damages resulting from the use of these calculations.