
The 15 Algorithms Machine Learning Engineers Need to Know
Machine learning has progressively increased greater notoriety in the recent years.
Probably the most well-known cases of machine learning are Facebook’s Algorithms to make movie proposals in light of films you have viewed in the past or Amazon’s Algorithms that suggest books in light of books you have purchased sometime recently Or Netflix’s suggestions on movies watched recently Or Youtube suggestions etc.
5 steps in the Machine Learning Process
Main Techniques of Machine Learning are:
- Classification
- Regression
- Clustering
- Recommendation Systems
- Anomaly Detection
- Dimensionality Reduction
3 Types of Machine Learning Algorithms:
Machine learning Algos can be isolated into 3 general categories :
1. Supervised learning,
2. Unsupervised learning and
3. Reinforcement learning.
Supervised learning is valuable in situations where a property (label) is accessible for a specific dataset ( training set), however, is missing and should be anticipated for other instances.
Unsupervised learning is helpful in situations where the test is to find certain connections in a given unlabeled dataset (things which are not pre-assigned).
Reinforcement learning falls between these, there is some type of input accessible for each prescient stride or activity, yet no exact name or blunder message.
Every one of the 3 strategies are utilized as a part of this list of 15 Basic Machine Learning Algorithms:
Infographic Credit : dataiku
1.) Linear regression
Linear regression is a linear system and the coefficients can be calculated analytically using linear algebra. … Linear regression does provide a useful exercise for learning stochastic gradient descent which is an important algorithm used for minimizing cost functions by machine learning algorithms.
2.) Logistic Regression
Logistic regression is another technique borrowed by machine learning from the field of statistics. It is the go-to method for binary classification problems (problems with two class values). … Techniques used to learn the coefficients of a logistic regression model from data.
3.) Decision Tree
Decision tree learning uses a decision tree (as a predictive model) to go from observations about an item (represented in the branches) to conclusions about the item’s target value (represented in the leaves). It is one of the predictive modelling approaches used in statistics, data mining and machine learning.
4.) SVM (Support Vector Machine)
In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data used for classification and regression analysis.
5.) Naive Bayes
In machine learning, naive Bayes classifiers are a family of simple probabilistic classifiers based on applying Bayes’ theorem with strong (naive) independence assumptions between the features. Naive Bayes has been studied extensively since the 1950s.
6.) Ensemble Methods
Stacking Multiple Machine Learning Models. Stacking, also known as stacked generalization, is an ensemble method where the models are combined using another machine learning algorithm. … Then this new dataset is used as input for the combiner machine learning algorithm.
7.) Clustering Algorithms
Clustering is a method of unsupervised learning and a common technique for statistical data analysis used in many fields. K-means clustering is an algorithm to classify or to group your objects based on attributes/features into K number of the group. K is a positive integer number.
8.) Principal Component Analysis
Principal component analysis (PCA) is a statistical procedure that uses an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of linearly uncorrelated variables called principal components (or sometimes, principal modes of variation).
9.) Singular Value Decomposition
In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It is the generalization of the eigendecomposition of a positive semidefinite normal matrix (for example, a symmetric matrix with positive eigenvalues) to any m × n {\displaystyle m\times n} matrix via an extension of the polar decomposition. It has many useful applications in signal processing and statistics.
10.) Independent Component Analysis
Independent component analysis attempts to decompose a multivariate signal into independent non-Gaussian signals. As an example, the sound is usually a signal that is composed of the numerical addition, at each time t, of signals from several sources. The question then is whether it is possible to separate these contributing sources from the observed total signal. When the statistical independence assumption is correct, blind ICA separation of a mixed signal gives very good results.It is also used for signals that are not supposed to be generated by a mixing for analysis purposes.
A simple application of ICA is the “cocktail party problem”, where the underlying speech signals are separated from a sample data consisting of people talking simultaneously in a room. Usually, the problem is simplified by assuming no time delays or echoes.
11.) KNN (K- Nearest Neighbors)
k-nearest neighbors algorithm. In pattern recognition, the k-nearest neighbors algorithm (k-NN) is a non-parametric method used for classification and regression. In both cases, the input consists of the k closest training examples in the feature space.
12.) Random Forest
Random forests or random decision forests are an ensemble learning method for classification, regression, and other tasks, that operate by constructing a multitude of decision trees at training time and outputting the class that is the mode of the classes (classification) or mean prediction (regression) of the individual trees. Random decision forests correct for decision trees’ habit of over-fitting to their training set.
13.) K-Means
k-means clustering is a method of vector quantization, originally from signal processing, that is popular for cluster analysis in data mining.
14.) Dimensionality Reduction Algorithms
In machine learning and statistics, dimensionality reduction or dimension reduction is the process of reducing the number of random variables under consideration, via obtaining a set of principal variables. It can be divided into feature selection and feature extraction
15.) Gradient Boosting & AdaBoost
Gradient boosting. Gradient boosting is a machine learning technique for regression and classification problems, which produces a prediction model in the form of an ensemble of weak prediction models, typically decision trees.
Conclusion:
In the event that you need to build a profession in machine learning, begin immediately. The field is developing rapidly, and the sooner you comprehend the extent of machine learning devices, the sooner you’ll have the capacity to give answers for complex work issues.
In the wake of understanding this 15 Algorithms you can begin making machine learning applications that improve encounters for individuals all over the place.
Related:
HOW TO LEARN MACHINE LEARNING IN 90 DAYS
6 Easy Steps To Get Started Learning Artificial Intelligence
10 Free Training Courses on Machine Learning and Artificial Intelligence
Very informative article
Thanks.
Keep visiting http://www.favouriteblog.com for new articles.