Every day more and more use cases are found for machine learning. It is a great field to get into.
We just released a 10-hour machine learning course for beginners on the freeCodeCamp.org YouTube channel.
Ayush Singh developed this course. He is a young data scientist and machine learning engineer.
Here are the sections covered in this course:
Section 1: Basics of Machine Learning
- What is Machine Learning? The way I like to think about it!
- Cool Applications of Machine Learning
- Types of ML and Their Types
- Workflow of basic ML Problem
- Main Challenges of Machine Learning
- Dividing the data
- Two famous problems of Machine learning: Underfitting and Overfitting
- Solutions to the Overfitting and Underfitting
- Supervised Learning and Unsupervised Learning In Depth
Section 2: Linear Regression & Regularization
- What is Linear Regression? Visual Understanding
- Hypothesis Function Or Prediction Function
- Closed Form Solution aka Normal Equation
- Coding Normal Equation
- Cost Function
- Gradient Descent
- Assumptions & Pros and Cons of Linear Regression
- Regularized Linear Models
- Ridge Regression
- Lasso Regression
Section 3: Logistic Regression & Performance Metrics
- Logistic Regression
- Hypothesis Function
- Cost Function
- Gradient Descent
- Assumptions and Pros and Cons
Section 4: Support Vector Machine
- Support Vector Machines
- Linear SVM Classification
- Hard/Soft Margin Classification
- Non-Linear SVM Classification
- Polynomial Kernel [Homogenous & Inhomogeneous ]
- RBF Kernel
- Computing SVM Classifier
- Primal and Dual Problem
- Sub-Gradient Descent
- Coordinate Descent
- Transductive SVM
- SVR
Section 5: PCA
- Review of Linear Transformation & EigenVectors and EigenValues
- Dimensionality Reduction Need
- Basic Intuition Behind PCA
- Data Preprocessing [Data Standardization]
- Compute the Covariance Matrix
- Compute the cumulative energy content for each eigenvector
- Select a subset of the eigenvectors as basis vectors
- Projecting Back
Section 6: Learning Theory
- Bias and Variance TradeOff
- Approx Estimation Error
- Empirical Risk Minimization
- Problem Sets releases
Section 7: Decision Trees & Random Forest
- Decision Trees
- Training of Decision Trees
- Prediction in Decision Trees
- Entropy
- Information Gain
- Gini Impurity
- Hyperparameter Tuning
- Project Proposal
- Decision Trees Assignment
- Ensemble Learning
- Ensemble Learning
- Bagging
- Random Forest
- Boosting
- Gradient Boosting
- Adaboost
- XGboost
- Stacking
- Cascading
Section 7.5: Learning more algorithms and building more projects
- Naive Bayes
- K-Nearest Neighbors
Section 8: Unsupervised Learning Algorithms
- Unsupervised Learning and Clustering Intro and Types
- Cluster Analysis
- Unsupervised Learning Algorithms
- K-Means with K-Means ++
- Hierarchical Clustering Techniques
- Unsupervised Learning Problem set and project releases
- K-Means Programming Assignment
Section 9: Building Applications
- Building Heart Failure Detection System with deployment
- Building Fake news detection system
- Building Email Spam Detection System
Watch the full course below or on the freeCodeCamp.org YouTube channel (10-hour watch).