Softmax function
Smooth approximation of one-hot arg max / From Wikipedia, the free encyclopedia
Dear Wikiwand AI, let's keep it short by simply answering these key questions:
Can you list the top facts and stats about Softmax?
Summarize this article for a 10 year old
SHOW ALL QUESTIONS
This article is about the smooth approximation of one-hot arg max. For the smooth approximation of max, see LogSumExp.
"Softmax" redirects here. For the Korean video game company, see ESA (company).
The softmax function, also known as softargmax[1]: 184 or normalized exponential function,[2]: 198 converts a vector of K real numbers into a probability distribution of K possible outcomes. It is a generalization of the logistic function to multiple dimensions, and used in multinomial logistic regression. The softmax function is often used as the last activation function of a neural network to normalize the output of a network to a probability distribution over predicted output classes.