Skip to content

Implementation and analysis of core Machine Learning Algorithms from scratch.

Notifications You must be signed in to change notification settings

ioangatop/MachineLearning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

38 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Machine Learning

License

Description

Welcome to our Machine Learning reposetory! Here you will find various projects from polynomial regression to fully-connected neural networks from scratch, SVM and Gaussian Processes! On Part II, we analize Independent Component Analysis, Graphical Models, EM and VAEs!

Machine Learning: Part I

Part 1: Polynomial Regression

Polynomial regressions as prediction function, along with the data and the original sine function of various polynomial order.

Part 2: Bayesian Linear (Polynomial) Regression

Bayesian linear regression model Left: Plot of predictive distribution Right: 100 polynomials sampled from the parameter posterior distribution

Part 1: Multiclass logistic regression

MNIST: Left: visualisation of the first 8 digits of the trainingset Right: visualization of leanred weights.

Left: Easiest digits for classification. Right: Hardest digits for classification

Part 2: Multilayer perceptron

Weights of the hidden layer at epoch 0, 4 and 9.

Comparison

Left: Multiclass logistic regression Right: Multilayer perceptron

Part 1: Gaussian Processes

Gaussian Processes.

Part 2: Support Vector Machines

Support Vector Machines.

Machine Learning: Part II

In this assignment, we implement the Independent Component Analysis algorithm, as described in chapter 34 of David MacKay's book "Information Theory, Inference, and Learning Algorithms".

Results of signal reconstruction using different priors and W matrix initialization.

In this assignment, we implement the sum-product and max-sum algorithms for factor graphs over discrete variables. We implemented these algorithms to a medical graph, in order to infer the possible decease.

Medical Directed Graph.

In this assignment, we implement the Expectation Maximization (EM) algorithm and Variational Autoencoder (VAE) on the MNIST dataset of written digits.

VAE's leanred manifold of the MNIST dataset of written digits.

Acknowledgement - References

The majority of the projects come from the lab assignments of the Machine Learning 1 and Machine Learning 2 courses of the MSc in Artificial Intelligence at the University of Amsterdam.