Skip to content

victor-explore/Machine-Learning-Notes-IISC-Banglore

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Machine Learning Notes

This repository contains comprehensive notes on various machine learning topics. These notes cover a wide range of concepts from basic principles to advanced algorithms.

Table of Contents

  1. Hypothesis Space
  2. Bayes Classifier
  3. Linear Regression
  4. Generalized Linear Regression
  5. Non-parametric Density Estimation
  6. Parzen Window Estimate
  7. K-Nearest Neighbour (KNN)
  8. Linear Discriminant Analysis (LDA)
  9. Support Vector Machine (SVM)
  10. Neural Networks
  11. Backpropagation
  12. Decision Trees
  13. Ensemble Learning
  14. Bagging and Random Forest
  15. Boosting
  16. XGBoost
  17. Principal Component Analysis (PCA)
  18. K-means Clustering
  19. Expectation Maximization (EM) Algorithm
  20. Miscellaneous Machine Learning Terms

Hypothesis Space

The hypothesis space is defined as the set of all possible hypothesis functions that map feature vectors to labels. It's represented as:

H = {h : X → Y}

More details

Bayes Classifier

The Bayes classifier is defined as:

h*(x) = argmax[y ∈ Y] P(Y = y | X = x)

It's proven to be the best classifier for the 0-1 loss function.

More details

Linear Regression

Linear regression models the relationship between input features and output as a linear function. The notes cover the formulation, ideal regressor derivation, and the closed-form solution.

More details

Generalized Linear Regression

This extends linear regression by projecting data into a higher-dimensional space before performing linear regression, allowing for capture of more complex relationships.

More details

Non-parametric Density Estimation

Non-parametric density estimation techniques estimate the probability density function directly from the data without assuming a specific functional form.

More details

Parzen Window Estimate

Also known as kernel density estimation, this method uses a window function to estimate the probability density function.

More details

K-Nearest Neighbour (KNN)

KNN is a non-parametric method used for classification and regression. The algorithm and its formulation are explained in detail.

More details

Linear Discriminant Analysis (LDA)

LDA is explained from a Bayesian perspective, including the derivation of the decision boundary.

More details

Support Vector Machine (SVM)

The notes cover SVM for both linearly separable and non-linearly separable data, as well as the kernel trick for handling non-linear decision boundaries.

More details

Neural Networks

The notes provide a mathematical formulation of neural networks and explain the importance of non-linear activation functions.

More details

Backpropagation

A detailed derivation of the backpropagation algorithm used for training neural networks is provided.

More details

Decision Trees

The notes cover how decision trees work, including the growing and pruning processes, and metrics like Gini Impurity and Mean Squared Error.

More details

Ensemble Learning

An introduction to ensemble learning techniques, which combine multiple models to improve overall performance.

More details

Bagging and Random Forest

Bagging (Bootstrap Aggregating) and Random Forest, which is an application of bagging to decision trees, are explained.

More details

Boosting

The notes provide a mathematical formulation of boosting, explaining how it sequentially trains models to correct errors of previous ones.

More details

XGBoost

XGBoost, a specific implementation of gradient boosting, is explained in detail.

More details

Principal Component Analysis (PCA)

PCA, a dimensionality reduction technique, is explained step-by-step, including the intuition behind it.

More details

K-means Clustering

K-means clustering, an unsupervised learning algorithm, is explained along with its connection to the Expectation-Maximization algorithm.

More details

Expectation Maximization (EM) Algorithm

The EM algorithm, used for finding maximum likelihood estimates of parameters in statistical models with latent variables, is derived and explained.

More details

Miscellaneous Machine Learning Terms

Various important machine learning terms and concepts are explained, including epochs, batch size, gradient descent variants, batch normalization, layer normalization, dropout, and N-fold cross-validation.

More details

Contributing

Contributions to improve or expand these notes are welcome. Please feel free to submit a pull request or open an issue for discussion.

License

This project is licensed under the MIT License - see the LICENSE.md file for details.

About

Notes of PRNN course taught at IISC as part of MTech AI curriculum

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors