Mathematics for machine learning
Introduction to linear algebra
Linear algebra is important in machine learning. It is in the heart
of many machine learning algorithms such as Singluar Value Decomposition (SVD) and
Principal Component Analysis (PCA). These are just few examples where linear algebra shines.
If you want to know how dimensionality reduction
works you ought to understand what matrices and vectors are, what are eigenvalues and eigenvectors,
what inverse of a matrix stands for, what is covariance matrix is, and many other
important aspects.
A very brief introduction (in just 4 pages) into linear algebra can be found
here. There exists a rife number
of literature on linear algebra.
Gradient descent
Gradient descent is a powerful algorithm that is being used in machine learning
extensively. In this article we are going to demonstrate the power of this algorithm
using simple example - polynomial regression. Keep on reading here.
Basics of Principal Component Analysis (PCA)
If you face a problem of large datasets and how to store those on disk
and later on use them to train the machine learning algorithms, dimensionality reduction
can help a lot. PCA allows one to reduce the amount of features in the dataset
with minimal lose of information (well at the end it is the engineer who controls how much information
he or she is eager to lose during the transformation).
Read more to discover how PCA works in practice.
Approximating matrices with Singular Value Decomposition (SVD)
SVD can be viewed as generalization to PCA. Similarly to PCA, SVD allows compressing
the matrices with minimal loss of information (well, one can control how much compression he
or she wants). In this article we show how to decompose
matrices using SVD.
Large margin classifier
Understanding the mathematics behind algorithms is important for several reasons.
One such reason is that the reader will have more clear picture of what the
algorithm is doing under the hood and how to fine tune it. In this article, we will discuss
the mathematics behind Support Vector Machine (SVM).
Math behind Artificial Neural Networks (ANN)
Artificial neural networks (or ANN) gained a lot attention in the last several years.
Development of specialized hardware and availability of various libraries make
this type of supervised machine learning algorithms indispensable tools in solving
challenging classification and regression problems. In this article
we will describe the mathematics behind ANN using simple example.
Utilizing conditional probabilities in classification problems
Naive Bayes classifier is the simple approach to classify categorical and continuous
data using conditional probabilities under the assumption that these probabilities
are independent for different features in the dataset. In this article we discuss
how to use this type of classifier.
Interpolating in space: Kriging
Often times there is a need to interpolate values which are spatially
scattered. In this cases we can use methods like inverse distance
weighting. However, this method does not take into consideration the
continuity of the data (direction), but only distance between
the point that we need to estimate and the rest of the known points.
To take into consideration both distance and direction,
it is best to use Kriging model. In this
article we are going to discuss the mathematics of simple Kriging
model.