Multi-output/
Multi-label Regression
Dr. Ashish Tendulkar
Machine Learning Techniques
IIT Madras
In case of multi-output regression, there are more than one output labels, all of which are real numbers.
Multi-output/ Multi-label Regression
Examples:
- Predicting number of runs scored by a batsman in next 5 innings.
- Daily average temperature of next 15 days.
Training Data
where
- \(\mathbf{X}\) is a \( n \times m\) input matrix.
- \(\mathbf{Y}\) is a \( n \times k\) label matrix.
- \(\mathbf{y}^{(i)} \in \mathcal{R}^k\) where \(k\) is the number of output labels, i.e., \(\mathbf{y}^{(i)}\) has \(k\) components that are real numbers.
- \(\mathbf{x}^{(i)}\) is a \(m \times 1\) feature vector
\( \mathbf{D} = (\mathbf{X}, \mathbf{Y}) = \{\left(\mathbf{x}^{(i)}, \mathbf{y}^{(i)}\right) \}_{i=0}^{n}\)
- The output label is a matrix \(\mathbf{Y}\). In order to generate multiple outputs, we need one weight vector per output.
- Hence, total of \(k\) weight vectors corresponding to the \(k\) outputs.
Model
\mathbf{Y}_{n \times k} = \mathbf{X}_{n \times (m+1)} \mathbf{W}_{(m+1) \times k}
There are two options for modeling this problem:
- Solve \(k\) independent linear regression problems. Gives some flexibility in using different representation for each problem.
- Solve a joint learning problem as outlined in the equation above.
Model
\mathbf{Y}_{n \times k} = \mathbf{X}_{n \times (m+1)} \mathbf{W}_{(m+1) \times k}
Loss
- Sum of squared error
J(\mathbf{W}) = \frac{1}{2}(\mathbf{X} \mathbf{W} - \mathbf{Y})^T (\mathbf{X} \mathbf{W} - \mathbf{Y})
Optimization
- Normal equation
- Gradient descent and its variations
Multi-output/Multi-label Regression
By Swarnim POD
Multi-output/Multi-label Regression
- 113