Machine Learning Algorithms - A Review: January 2019
Machine Learning Algorithms - A Review: January 2019
Machine Learning Algorithms - A Review: January 2019
net/publication/344717762
CITATIONS READS
16 37,546
1 author:
Batta Mahesh
Independent Researcher
5 PUBLICATIONS 16 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Batta Mahesh on 17 October 2020.
Abstract: Machine learning (ML) is the scientific study of algorithms and statistical models that computer systems use to perform a
specific task without being explicitly programmed. Learning algorithms in many applications that’s we make use of daily. Every time a
web search engine like Google is used to search the internet, one of the reasons that work so well is because a learning algorithm that
has learned how to rank web pages. These algorithms are used for various purposes like data mining, image processing, predictive
analytics, etc. to name a few. The main advantage of using machine learning is that, once an algorithm learns what to do with data, it
can do its work automatically. In this paper, a brief review and future prospect of the vast applications of machine learning algorithms
has been made.
Keywords: Algorithm, Machine Learning, Pseudo Code, Supervised learning, Unsupervised learning, Reinforcement learning
Machine Learning relies on different algorithms to solve learn some kind of patterns from the training dataset and
data problems. Data scientists like to point out that there‟s apply them to the test dataset for prediction or classification.
no single one-size-fits-all type of algorithm that is best to The workflow of supervised machine learning algorithms is
solve a problem. The kind of algorithm employed depends given in fig below. Most famous supervised machine
on the kind of problem you wish to solve, the number of learning algorithms have been discussed here
variables, the kind of model that would suit it best and so on.
Here‟s a quick look at some of the commonly used
algorithms in machine learning (ML)
Supervised Learning
Supervised learning is the machine learning task of learning
a function that maps an input to an output based on example
input-output pairs. It infers a function from labelled training
data consisting of a set of training examples. The supervised
machine learning algorithms are those algorithms which Figure: Supervised learning Workflow
needs external assistance. The input dataset is divided into
train and test dataset. The train dataset has output variable
which needs to be predicted or classified. All algorithms
Volume 9 Issue 1, January 2020
www.ijsr.net
Licensed Under Creative Commons Attribution CC BY
Paper ID: ART20203995 DOI: 10.21275/ART20203995 381
International Journal of Science and Research (IJSR)
ISSN: 2319-7064
ResearchGate Impact Factor (2018): 0.28 | SJIF (2018): 7.426
Decision Tree
Decision tree is a graph to represent choices and their results
in form of a tree. The nodes in the graph represent an event
or choice and the edges of the graph represent the decision
rules or conditions. Each tree consists of nodes and
branches. Each node represents attributes in a group that is
to be classified and each branch represents a value that the
node can take.
Navie Bayes
It is a classification technique based on Bayes Theorem with
an assumption of independence among predictors. In simple
terms, a Naive Bayes classifier assumes that the presence of
a particular feature in a class is unrelated to the presence of
any other feature. Naïve Bayes mainly targets the text
classification industry. It is mainly used for clustering and
classification purpose depends on the conditional probability
of happening.
Figure: Support Vector Machine
K-Means Clustering The next step is to take each point belonging to a given data
K-means is one of the simplest unsupervised learning set and associate it to the nearest center. When no point is
algorithms that solve the well known clustering problem. pending, the first step is completed and an early group age
The procedure follows a simple and easy way to classify a is done. At this point we need to re-calculate k new centroids
given data set through a certain number of clusters. The as bary center of the clusters resulting from the previous
main idea is to define k centers, one for each cluster. These step.
centers should be placed in a cunning way because of
different location causes different result. So, the better
choice is to place them is much as possible far away from
each other.
Self-Training
In self-training, a classifier is trained with a portion of
labeled data. The classifier is then fed with unlabeled data.
The unlabeled points and the predicted labels are added
together in the training set. This procedure is then repeated
further. Since the classifier is learning itself, hence the name
self-training.
Reinforcement Learning
Reinforcement learning is an area of machine learning
concerned with how software agents ought to take actions in
an environment in order to maximize some notion of
cumulative reward. Reinforcement learning is one of three
basic machine learning paradigms, alongside supervised
learning and unsupervised learning.
Bagging
Bagging or bootstrap aggregating is applied where the
accuracy and stability of a machine learning algorithm needs
to be increased. It is applicable in classification and
Figure: Reinforcement Learning regression. Bagging also decreases variance and helps in
handling overfitting.
Multitask Learning
Multi-Task learning is a sub-field of Machine Learning that
aims to solve multiple different tasks at the same time, by
taking advantage of the similarities between different tasks.
This can improve the learning efficiency and also act as a
regularize. Formally, if there are n tasks (conventional deep
Neural Networks
A neural network is a series of algorithms that endeavors to
recognize underlying relationships in a set of data through a Figure: Unsupervised Neural Network
process that mimics the way the human brain operates. In
this sense, neural networks refer to systems of neurons, Reinforced Neural Network
either organic or artificial in nature. Neural networks can Reinforcement learning refers to goal-oriented algorithms,
adapt to changing input; so the network generates the best which learn how to attain a complex objective (goal) or
possible result without needing to redesign the output maximize along a particular dimension over many steps; for
criteria. The concept of neural networks, which has its roots example, maximize the points won in a game over many
in artificial intelligence, is swiftly gaining popularity in the moves. They can start from a blank slate, and under the right
development of trading systems. conditions they achieve superhuman performance. Like a
child incentivized by spankings and candy, these algorithms
are penalized when they make the wrong decisions and
rewarded when they make the right ones – this is
reinforcement.
K-Nearest Neighbor
The k-nearest neighbors (KNN) algorithm is a simple,
supervised machine learning algorithm that can be used to
Figure: Supervised Neural Network solve both classification and regression problems. It's easy to
implement and understand, but has a major drawback of
2. Conclusion
Machine Learning can be a Supervised or Unsupervised. If
you have lesser amount of data and clearly labelled data for
training, opt for Supervised Learning. Unsupervised
Learning would generally give better performance and
results for large data sets. If you have a huge data set easily
available, go for deep learning techniques. You also have
learned Reinforcement Learning and Deep Reinforcement
Learning. You now know what Neural Networks are, their
applications and limitations. This paper surveys various
machine learning algorithms. Today each and every person
is using machine learning knowingly or unknowingly. From
getting a recommended product in online shopping to
updating photos in social networking sites. This paper gives
an introduction to most of the popular machine learning
algorithms.
References
[1] W. Richert, L. P. Coelho, “Building Machine Learning
Systems with Python”, Packt Publishing Ltd., ISBN
978-1-78216-140-0
[2] J. M. Keller, M. R. Gray, J. A. Givens Jr., “A Fuzzy K-
Nearest Neighbor Algorithm”, IEEE Transactions on
Systems, Man and Cybernetics, Vol. SMC-15, No. 4,
August 1985
[3] https://www.geeksforgeeks.org/machine-learning/
[4] ] S. Marsland, Machine learning: an algorithmic
perspective. CRC press, 2015.
[5] M. Bkassiny, Y. Li, and S. K. Jayaweera, “A survey on
machine learning techniques in cognitive radios,” IEEE
Communications Surveys & Tutorials, vol. 15, no. 3,
pp. 1136–1159, Oct. 2012.
[6] https://en.wikipedia.org/wiki/Instance-based_learning
[7] R. S. Sutton, “Introduction: The Challenge of
Reinforcement Learning”, Machine Learning, 8, Page
225-227, Kluwer Academic Publishers, Boston, 1992
[8] P. Harrington, “Machine Learning in action”, Manning
Publications Co., Shelter Island, New York, 2012