Softmax Regression (Multinomial Logistic Regression) / Softmax Multi-Class Network / Softmax Classifier - Python and Machine Learning for Integrated Circuits - - An Online Book - |
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Python and Machine Learning for Integrated Circuits http://www.globalsino.com/ICs/ | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Chapter/Index: Introduction | A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z | Appendix | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
================================================================================= Softmax regression, also known as multinomial logistic regression, softmax multi-class network, or a softmax classifier, is a machine learning algorithm used for classification tasks. It is an extension of logistic regression that is used when the task involves classifying instances into more than two classes. In softmax regression, the goal is to assign a probability to each class for a given input and then predict the class with the highest probability. It is particularly common in machine learning and deep learning applications for tasks like image classification, natural language processing, and more. Softmax regression is considered another member of the Generalized Linear Model (GLM) family because it shares several key characteristics and principles with GLMs. Generalized Linear Models are a broad class of statistical models used for various types of regression and classification tasks, and softmax regression can be viewed as a specific instance within this family. A oftmax multi-class network works in the way below:
---------------------------- [3854a] where, is the raw score (logit) for class is the total number of classes. Cross-entropy is commonly used as the loss function for training Softmax regression models:
The decision boundaries in Figure 3854 represent hyperplanes. In a real Softmax Regression model, these boundaries would be learned from the data.
Softmax function is commonly used in neural networks, especially in the output layer for classification problems. In neural networks, the softmax function is often used to convert a vector of raw scores (also called logits) into a probability distribution. The softmax function takes an input vector, exponentiates each element, and then normalizes the results so that they sum to 1. This is useful in classification problems where you want the network to output probabilities for each class. The softmax function is defined for a vector with elements as follows:----------------------------------------------- [3854a] where,:
The division by the sum ensures that the resulting values form a valid probability distribution. The goal with the softmax function is to convert these logits into probabilities that sum to 1. For the example described in page3876, with softmax function, then we can have softmax multi-class network below,
In this case, the division by the sum ensures that the resulting values form a valid probability distribution. The goal with the softmax function is to convert these logits into probabilities that sum to 1. Therefore, the probabilities of the three animals depend on each other. The softmax multi-class network is a type of neural network architecture used for classification tasks where the goal is to assign an input into one of multiple classes. ============================================
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
================================================================================= | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||