From 337301edba25ed6ac4fba71965ace6b570f07851 Mon Sep 17 00:00:00 2001 From: Julian T Date: Wed, 8 Jun 2022 17:12:49 +0200 Subject: Begin writing notes for exam --- sem8/ml/exam.tex | 222 +++++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 222 insertions(+) create mode 100644 sem8/ml/exam.tex (limited to 'sem8') diff --git a/sem8/ml/exam.tex b/sem8/ml/exam.tex new file mode 100644 index 0000000..0ba71e3 --- /dev/null +++ b/sem8/ml/exam.tex @@ -0,0 +1,222 @@ +\title{Exam Notes} + +\section{Linear Models} + +\subsection{K Nearest Neighbour} + +Near neighbours tend to have the same label. +Therefore prediction is done with the label that occurs most frequently among the $k$ nearest neighbours to the node itself. +Here distance is in accordance to a $D$-dimensional input feature space. + +\begin{mdframed}[nobreak,frametitle={A Note on Features}] + \begin{itemize} + \item[$\mathbf{x}$] Information we get raw, such as \emph{attributes}, \emph{(input) features}, \emph{(predictor) variables}. + \item[$\mathbf{x'}$] Selected, transformed, or otherwise ``engineered'' features. + \item[$\mathbf{x''}$] A transformation of $x'$ that is automaticly constructed by the machine learning method. + \end{itemize} +\end{mdframed} + +\begin{description} + \item[Instance space] Space of all possible values of input features $\mathbf{x}$. + \item[Decision regions] A classifier devides the instance space into decision regions. +\end{description} + +\subsection{Perceptron and Naive Bayes} + +Linear models, which are therefore limited in the applicability (they cannot predict XOR function). +However they are usefull baselines, and can be trained well with limited datasets. +They are also integral components in \emph{support vector machines} and \emph{deep neural networks}. + +Less unlikely to overfit, however we can still apply techniques to prevent it. + +\subsubsection{The Perceptron} + +A neural network with a input layer, no hidden layers, and a single output neuron with a \emph{sign} activation function. +This is represented as, +\[ + O(x_1, \dots, x_n) = \left\{ + \begin{array}{ll} + 1 & \mathrm{if}\; w_0 + w_1 \cdot x_1 + \dots + w_n \cdot x_n > 0 \\ + -1 & \mathrm{otherwise} + \end{array} + \right. +\] + +This can be used to classify a binary class, where the decisions are seperated by a \emph{linear hyperplane}. +Therefore it is not possible to classify something like the XOR function. + +\subsubsection{Naive Bayes Model} + +Here we instead work with probabilities, as opposed to the perceptron. +It assumes that attributes are independent, given the class label. + +Prediction is done by comparing the probability of each label, and then choosing the most likely. +Thus if labels are $\oplus$ and $\oslash$, we choose $\oplus$ if +\[ + P(\oplus \mid X_1, ..., X_n) \geq P(\oslash \mid X_1, ..., X_n) +\] + +\subsection{Overfitting} + +Our hypothesis overfits the training data if there exists another hypothesis which performs worse in traning but better in testing. +Thus we learn our training data too well, thus not capturing the general characteristics of what we are predicting. + +This can be the case when the \emph{hypothesis space} is very large, refering to the complexity of the learned structure. + +\subsection{Linear Functions and Decision Regions} + +Linear function can be written with scalar values or vectors, as follows: +\[ + y(x_1, ..., x_D) = w_0 + w_1 \cdot w_1 + \dots + w_D \cdot x_D = w_0 + \mathbf{w} \cdot \mathbf{x} +\] +Using it to make decisions, is done with \emph{decision regions}, +\begin{align*} + \mathcal{R}_1 &= \{\mathbf{x} \mid y(\mathbf{x}) \geq 0\} \\ + \mathcal{R}_2 &= \{\mathbf{x} \mid y(\mathbf{x}) < 0 \} +\end{align*} + +Look at the geometry, $\mathbf{w}$ is the direction of the decision boundary between the two binaries. +And $w_0 / || \mathbf{w} ||$, is the distance between the decision boundary and origin. + +\subsubsection{Multiple Classes} + +If more than two classes is wanted, one can use multiple linear functions in combination. +There are different approaches to this, and they have acompanying figures in the slides. +\begin{itemize} + \item Multiple binary "one against all" classification. + \item Multiple binary "one against one" classification. +\end{itemize} + +Instead one can construct a \emph{discriminant function} for each of the class labels. +Then we can classify some input, $\mathbf{x}$, as a label if the corrosponding function is maximal. + +\subsubsection{Least Squares Regresssion} + +For each data case $\mathbf{x}_n$ have a \emph{target vector}, where class labels are encoded with \emph{one-hot encoding}. +Then we try to minimize +\[ + E_D(\mathbf{\tilde{W}}) = \frac 1 2 \sum_{n=1}^N || \mathbf{\tilde{W}}^T \mathbf{\tilde{x}}_n - \mathbf{t}_n ||^2 +\] + +It should be noted that this method does not minimize classification errors, and outliers may cause a learned function to not seperate linearly seperable data correctly. + +\subsection{Probabilistic Models} + +Here we classify $\mathbf{x}$ as class $k$ if $P(Y = k \mid \mathbf{x})$ is maximal. +There are two approaches. + +\paragraph{Generative Approach} Where we classify from simple assumptions about the distribution of the data. + Thus we model the probabilities $P(\mathbf{x} \mid k)$ (class-conditional) and $P(k)$ (class prior) and from this calculate $P(k \mid \mathbf{x})$ (posterior). + Examples of this include \emph{Naive Bayes} and \emph{LDA}. + Has the consequence that we can generate example data. + + When learning, the goal is to maximize the likelihood: + \[ + \prod_{n=1}^N P(\mathbf{x}_n, y_n) + \] + +\paragraph{Discriminative Approach} Here we directly learn the distribution $P(Y \mid \mathbf{X})$. + Examples are \emph{logistic regression} and \emph{neural networks}. + + Likewise the learning goal is to maximize the likelihood: + \[ + \prod_{n=1}^N P(y_n \mid \mathbf{x}_n) + \] + Note that this is a conditional probablity and not a joint probability. + +\subsubsection{Naive Bayes} + +\begin{figure}[H] + \centering + \begin{tikzpicture} + \node[draw, circle] (y) {$\;Y\;$}; + \node[draw, circle, below=of y] (x2) {$X_2$}; + \node[draw, circle, right=of x2] (x3) {$X_3$}; + \node[draw, circle, left=of x2] (x1) {$X_1$}; + + \draw[->] (y) edge (x1) (y) edge (x2) (y) edge (x3); + + \end{tikzpicture} +\end{figure} + +Here $P(Y \mid \mathbf{x})$ is a probability table for each class. +And $P(X \mid Y, \mathbf{w})$ is if $X_i$ is categorical a table, otherwise multiple gaussian distributions with parameters defined by $\mathbf{w}$. + +\subsubsection{Gaussian Mixture Models} + +Here we do not have the assumption that input variables are independent as with the naive bayes method. +Instead we model $P(\mathbf{X} \mid Y)$ as Gaussian, such that for all classes $k$: +\[ + P(\mathbf{X} = \mathbf{x} \mid Y = y) = \frac 1 {(2\pi)^{D / 2} | \Sigma_k |^{1 / 2}} \cdot e^{-\frac 1 2 (\mathbf{x} - \mathbf{\mu}_k)^T \Sigma_k^{-1}(\mathbf{x} - \mathbf{\mu}_k)} +\] +defined with the parameters $\mathbf{\mu}_k$ (mean vectors) and $\Sigma_k$ (co-variance matrixes). + +\subsubsection{Logistic Regression} + +A model with continues input and two output classes: $0$ and $1$. +And yes this is classification and not regression (what). + +\subsection{Exam Notes} + +The topics for the exam are as follows, the missing topics are noted in fat: +\begin{itemize} + \item Decision Regions + \item Overfitting + \item Least squares regression (corresponding to sklearn LinearRegression in self study 1) + \item Linear discriminant analysis + \item \textbf{Logistic Regression} +\end{itemize} + +\begin{mdframed}[nobreak,frametitle={Questions that Need Answering}] + \begin{itemize} + \item What is the logistic regression. + \item Løs opgaver for denne lektion. + \end{itemize} +\end{mdframed} + +\subsubsection{Self Study 1} + +\begin{enumerate} + \item Dataset contains for \emph{features}. We show a plot of two of the features, where two of the three classes are \emph{linearly seperable}. + \item We create 4 models one of which is the linear regression. + \item We know note the \emph{decision regions}, some important points: + \begin{itemize} + \item KNN can have multiple decision regions for the same class, and the \emph{boundaries} are not linear. + This is clear when we have a $N=1$, and a green outlier has its own small region. + \item With linear regression, we use 3 \emph{decision functions} and then choose the class which has the largest output. + In this case, this does not work that well, given that least squares comes from regression which assumes a gaussian distribution which does not fit binary values. + \end{itemize} + \item Then we split the data using a seed of one, and predict with KNN ($N=1$), KNN ($N=3$), and linear regression. + We see that linear regression gives much worse results, which makes sense given the decision regions that where drawn. + From the \emph{confusion matrix} we se that green and blue are very confusing which makes sense. + \item We now try with all the features. + Here we see that KNN ($N=3$) gives the best results on the testing data, however actually performs worse than $N=1$ on the training data. + This may be due to the \emph{overfitting} which happens when increasing the complexity of the model structure. + + Linear also performs much better, due to it being able to use other features the linearly seperate classes. +\end{enumerate} + + +\section{Support Vector Machines} + +\subsection{Exam Notes} + +\begin{itemize} + \item Maximum margin hypeplanes + \item Feature transformations and kernel functions + \item The kernel trick + \item String kernels +\end{itemize} + +\subsection{Feature Space} + +We can create a mapping, $\phi : \mathbb{R}^D \rightarrow \mathbb{R}^{D'}$, which transforms the original data: +\[ + \phi(\mathbf{x}) = (\phi_1(\mathbf{x}), \dots, \phi_{D'}(\mathbf{x})) +\] +Here the components $\phi_i$ are called \emph{features} or \emph{basis functions} and $\mathbb{R}^{D'}$ is the \emph{feature space} of $\phi$. +This mapping can be usefull, to transform data such that it is linearly seperable. + +\subsection{Nonlinear SVM} + + -- cgit v1.2.3