ml_cheatsheet
ml_cheatsheet
Table of Contents
1. Supervised Learning
Linear Regression
Logistic Regression
Decision Trees
Random Forests
Support Vector Machines (SVM)
k-Nearest Neighbors (k-NN)
Naive Bayes
Gradient Boosting Machines (GBM)
Neural Networks
2. Unsupervised Learning
k-Means Clustering
Hierarchical Clustering
Principal Component Analysis (PCA)
Independent Component Analysis (ICA)
Association Rules
Autoencoders
3. Reinforcement Learning
Q-Learning
Deep Q-Networks (DQN)
Policy Gradients
Actor-Critic Methods
4. Semi-Supervised and Self-Supervised Learning
Self-Training
Co-Training
5. Ensemble Methods
Bagging
Boosting
Stacking
Supervised Learning
1. Linear Regression
2. Logistic Regression
3. Decision Trees
4. Random Forests
7. Naive Bayes
Purpose: Classification.
Key Concept: Based on Bayes' Theorem with the assumption of feature independence.
Variants: Gaussian, Multinomial, Bernoulli.
9. Neural Networks
Unsupervised Learning
1. k-Means Clustering
2. Hierarchical Clustering
5. Association Rules
6. Autoencoders
Reinforcement Learning
1. Q-Learning
3. Policy Gradients
4. Actor-Critic Methods
1. Self-Training
2. Co-Training
Ensemble Methods
2. Boosting
3. Stacking
Elastic Net
Overfitting: Model performs well on training data but poorly on unseen data.
Underfitting: Model is too simple to capture underlying patterns.
Bias-Variance Tradeoff: Balance between model complexity and generalization.
Cross-Validation: Technique to assess model performance by partitioning data.
Regularization: Techniques to prevent overfitting (e.g., L1, L2).
Feature Scaling: Standardizing features to improve model performance.
Python Libraries:
Books:
"Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow" by Aurélien Géron
"Pattern Recognition and Machine Learning" by Christopher M. Bishop
"The Elements of Statistical Learning" by Trevor Hastie, Robert Tibshirani, and Jerome Friedman
Online Courses: