13 min read

An Introduction to Machine Learning Theory and Its Applications: A Visual Tutorial with Examples

View all articles

Machine Learning (ML) is coming into its own, with a growing recognition that ML can play a key role in a wide range of critical applications, such as data mining, natural language processing, image recognition, and expert systems. ML provides potential solutions in all these domains and more, and is set to be a pillar of our future civilization.

The supply of able ML designers has yet to catch up to this demand. A major reason for this is that ML is just plain tricky. This Machine Learning tutorial introduces the basics of ML theory, laying down the common themes and concepts, making it easy to follow the logic and get comfortable with machine learning basics.

Machine learning tutorial illustration: This curious machine is learning machine learning, unsupervised.

What is Machine Learning?

So what exactly is “machine learning” anyway? ML is actually a lot of things. The field is quite vast and is expanding rapidly, being continually partitioned and sub-partitioned ad nauseam into different sub-specialties and types of machine learning.

There are some basic common threads, however, and the overarching theme is best summed up by this oft-quoted statement made by Arthur Samuel way back in 1959: “[Machine Learning is the] field of study that gives computers the ability to learn without being explicitly programmed.”

And more recently, in 1997, Tom Mitchell gave a “well-posed” definition that has proven more useful to engineering types: “A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E.”

“A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E.” -- Tom Mitchell, Carnegie Mellon University

So if you want your program to predict, for example, traffic patterns at a busy intersection (task T), you can run it through a machine learning algorithm with data about past traffic patterns (experience E) and, if it has successfully “learned”, it will then do better at predicting future traffic patterns (performance measure P).

The highly complex nature of many real-world problems, though, often means that inventing specialized algorithms that will solve them perfectly every time is impractical, if not impossible. Examples of machine learning problems include, “Is this cancer?”, “What is the market value of this house?”, “Which of these people are good friends with each other?”, “Will this rocket engine explode on take off?”, “Will this person like this movie?”, “Who is this?”, “What did you say?”, and “How do you fly this thing?”. All of these problems are excellent targets for an ML project, and in fact ML has been applied to each of them with great success.

ML solves problems that cannot be solved by numerical means alone.

Among the different types of ML tasks, a crucial distinction is drawn between supervised and unsupervised learning:

  • Supervised machine learning: The program is “trained” on a pre-defined set of “training examples”, which then facilitate its ability to reach an accurate conclusion when given new data.
  • Unsupervised machine learning: The program is given a bunch of data and must find patterns and relationships therein.

We will primarily focus on supervised learning here, but the end of the article includes a brief discussion of unsupervised learning with some links for those who are interested in pursuing the topic further.

Supervised Machine Learning

In the majority of supervised learning applications, the ultimate goal is to develop a finely tuned predictor function h(x) (sometimes called the “hypothesis”). “Learning” consists of using sophisticated mathematical algorithms to optimize this function so that, given input data x about a certain domain (say, square footage of a house), it will accurately predict some interesting value h(x) (say, market price for said house).

In practice, x almost always represents multiple data points. So, for example, a housing price predictor might take not only square-footage (x1) but also number of bedrooms (x2), number of bathrooms (x3), number of floors (x4), year built (x5), zip code (x6), and so forth. Determining which inputs to use is an important part of ML design. However, for the sake of explanation, it is easiest to assume a single input value is used.

So let’s say our simple predictor has this form:

h of x equals theta 0 plus theta 1 times x

where theta 0 and theta 1 are constants. Our goal is to find the perfect values of theta 0 and theta 1 to make our predictor work as well as possible.

Optimizing the predictor h(x) is done using training examples. For each training example, we have an input value x_train, for which a corresponding output, y, is known in advance. For each example, we find the difference between the known, correct value y, and our predicted value h(x_train). With enough training examples, these differences give us a useful way to measure the “wrongness” of h(x). We can then tweak h(x) by tweaking the values of theta 0 and theta 1 to make it “less wrong”. This process is repeated over and over until the system has converged on the best values for theta 0 and theta 1. In this way, the predictor becomes trained, and is ready to do some real-world predicting.

Machine Learning Examples

We stick to simple problems in this post for the sake of illustration, but the reason ML exists is because, in the real world, the problems are much more complex. On this flat screen we can draw you a picture of, at most, a three-dimensional data set, but ML problems commonly deal with data with millions of dimensions, and very complex predictor functions. ML solves problems that cannot be solved by numerical means alone.

With that in mind, let’s look at a simple example. Say we have the following training data, wherein company employees have rated their satisfaction on a scale of 1 to 100:

Employee satisfaction rating by salary is a great machine learning example.

First, notice that the data is a little noisy. That is, while we can see that there is a pattern to it (i.e. employee satisfaction tends to go up as salary goes up), it does not all fit neatly on a straight line. This will always be the case with real-world data (and we absolutely want to train our machine using real-world data!). So then how can we train a machine to perfectly predict an employee’s level of satisfaction? The answer, of course, is that we can’t. The goal of ML is never to make “perfect” guesses, because ML deals in domains where there is no such thing. The goal is to make guesses that are good enough to be useful.

It is somewhat reminiscent of the famous statement by British mathematician and professor of statistics George E. P. Box that “all models are wrong, but some are useful”.

The goal of ML is never to make “perfect” guesses, because ML deals in domains where there is no such thing. The goal is to make guesses that are good enough to be useful.

Machine Learning builds heavily on statistics. For example, when we train our machine to learn, we have to give it a statistically significant random sample as training data. If the training set is not random, we run the risk of the machine learning patterns that aren’t actually there. And if the training set is too small (see law of large numbers), we won’t learn enough and may even reach inaccurate conclusions. For example, attempting to predict company-wide satisfaction patterns based on data from upper management alone would likely be error-prone.

With this understanding, let’s give our machine the data we’ve been given above and have it learn it. First we have to initialize our predictor h(x) with some reasonable values of theta 0 and theta 1. Now our predictor looks like this when placed over our training set:

h of x equals twelve plus 0 point two x

Machine learning example illustration: A machine learning predictor over a training dataset.

If we ask this predictor for the satisfaction of an employee making $60k, it would predict a rating of 27:

In this image, the machine has yet to learn to predict a probable outcome.

It’s obvious that this was a terrible guess and that this machine doesn’t know very much.

So now, let’s give this predictor all the salaries from our training set, and take the differences between the resulting predicted satisfaction ratings and the actual satisfaction ratings of the corresponding employees. If we perform a little mathematical wizardry (which I will describe shortly), we can calculate, with very high certainty, that values of 13.12 for theta 0 and 0.61 for theta 1 are going to give us a better predictor.

h of x equals thirteen point one two plus 0 point six one x

In this case, the machine learning predictor is getting closer.

And if we repeat this process, say 1500 times, our predictor will end up looking like this:

h of x equals fifteen point five four plus 0 point seven five x

With a lot of repetition, the machine learning process starts to take shape.

At this point, if we repeat the process, we will find that theta 0 and theta 1 won’t change by any appreciable amount anymore and thus we see that the system has converged. If we haven’t made any mistakes, this means we’ve found the optimal predictor. Accordingly, if we now ask the machine again for the satisfaction rating of the employee who makes $60k, it will predict a rating of roughly 60.

In this example, the machine has learned to predict a probable data point.

Now we’re getting somewhere.

Machine Learning Regression: A Note on Complexity

The above example is technically a simple problem of univariate linear regression, which in reality can be solved by deriving a simple normal equation and skipping this “tuning” process altogether. However, consider a predictor that looks like this:

Four dimensional equation example

This function takes input in four dimensions and has a variety of polynomial terms. Deriving a normal equation for this function is a significant challenge. Many modern machine learning problems take thousands or even millions of dimensions of data to build predictions using hundreds of coefficients. Predicting how an organism’s genome will be expressed, or what the climate will be like in fifty years, are examples of such complex problems.

Many modern ML problems take thousands or even millions of dimensions of data to build predictions using hundreds of coefficients.

Fortunately, the iterative approach taken by ML systems is much more resilient in the face of such complexity. Instead of using brute force, a machine learning system “feels its way” to the answer. For big problems, this works much better. While this doesn’t mean that ML can solve all arbitrarily complex problems (it can’t), it does make for an incredibly flexible and powerful tool.

Gradient Descent - Minimizing “Wrongness”

Let’s take a closer look at how this iterative process works. In the above example, how do we make sure theta 0 and theta 1 are getting better with each step, and not worse? The answer lies in our “measurement of wrongness” alluded to previously, along with a little calculus.

The wrongness measure is known as the cost function (a.k.a., loss function), J of theta. The input theta represents all of the coefficients we are using in our predictor. So in our case, theta is really the pair theta 0 and theta 1. J of theta 0 and theta 1 gives us a mathematical measurement of how wrong our predictor is when it uses the given values of theta 0 and theta 1.

The choice of the cost function is another important piece of an ML program. In different contexts, being “wrong” can mean very different things. In our employee satisfaction example, the well-established standard is the linear least squares function:

Cost function expressed as a linear least squares function

With least squares, the penalty for a bad guess goes up quadratically with the difference between the guess and the correct answer, so it acts as a very “strict” measurement of wrongness. The cost function computes an average penalty over all of the training examples.

So now we see that our goal is to find theta 0 and theta 1 for our predictor h(x) such that our cost function J of theta 0 and theta 1 is as small as possible. We call on the power of calculus to accomplish this.

Consider the following plot of a cost function for some particular Machine Learning problem:

This graphic depicts the bowl-shaped plot of a cost function for a machine learning example.

Here we can see the cost associated with different values of theta 0 and theta 1. We can see the graph has a slight bowl to its shape. The bottom of the bowl represents the lowest cost our predictor can give us based on the given training data. The goal is to “roll down the hill”, and find theta 0 and theta 1 corresponding to this point.

This is where calculus comes in to this machine learning tutorial. For the sake of keeping this explanation manageable, I won’t write out the equations here, but essentially what we do is take the gradient of J of theta 0 and theta 1, which is the pair of derivatives of J of theta 0 and theta 1 (one over theta 0 and one over theta 1). The gradient will be different for every different value of theta 0 and theta 1, and tells us what the “slope of the hill is” and, in particular, “which way is down”, for these particular thetas. For example, when we plug our current values of theta into the gradient, it may tell us that adding a little to theta 0 and subtracting a little from theta 1 will take us in the direction of the cost function-valley floor. Therefore, we add a little to theta 0, and subtract a little from theta 1, and voilà! We have completed one round of our learning algorithm. Our updated predictor, h(x) = theta 0 + theta 1x, will return better predictions than before. Our machine is now a little bit smarter.

This process of alternating between calculating the current gradient, and updating the thetas from the results, is known as gradient descent.

This image depicts an example of a machine learning gradient descent.

This image depicts the number of iterations for this machine learning tutorial.

That covers the basic theory underlying the majority of supervised Machine Learning systems. But the basic concepts can be applied in a variety of different ways, depending on the problem at hand.

Classification Problems in Machine Learning

Under supervised ML, two major subcategories are:

  • Regression machine learning systems: Systems where the value being predicted falls somewhere on a continuous spectrum. These systems help us with questions of “How much?” or “How many?”.
  • Classification machine learning systems: Systems where we seek a yes-or-no prediction, such as “Is this tumer cancerous?”, “Does this cookie meet our quality standards?”, and so on.

As it turns out, the underlying Machine Learning theory is more or less the same. The major differences are the design of the predictor h(x) and the design of the cost function J of theta.

Our examples so far have focused on regression problems, so let’s now also take a look at a classification example.

Here are the results of a cookie quality testing study, where the training examples have all been labeled as either “good cookie” (y = 1) in blue or “bad cookie” (y = 0) in red.

This example shows how a machine learning regression predictor is not the right solution here.

In classification, a regression predictor is not very useful. What we usually want is a predictor that makes a guess somewhere between 0 and 1. In a cookie quality classifier, a prediction of 1 would represent a very confident guess that the cookie is perfect and utterly mouthwatering. A prediction of 0 represents high confidence that the cookie is an embarrassment to the cookie industry. Values falling within this range represent less confidence, so we might design our system such that prediction of 0.6 means “Man, that’s a tough call, but I’m gonna go with yes, you can sell that cookie,” while a value exactly in the middle, at 0.5, might represent complete uncertainty. This isn’t always how confidence is distributed in a classifier but it’s a very common design and works for purposes of our illustration.

It turns out there’s a nice function that captures this behavior well. It’s called the sigmoid function, g(z), and it looks something like this:

h of x equals g of z

The sigmoid function at work to accomplish a supervised machine learning example.

z is some representation of our inputs and coefficients, such as:

z equals theta 0 plus theta 1 times x

so that our predictor becomes:

h of x equals g of theta 0 plus theta 1 times x

Notice that the sigmoid function transforms our output into the range between 0 and 1.

The logic behind the design of the cost function is also different in classification. Again we ask “what does it mean for a guess to be wrong?” and this time a very good rule of thumb is that if the correct guess was 0 and we guessed 1, then we were completely and utterly wrong, and vice-versa. Since you can’t be more wrong than absolutely wrong, the penalty in this case is enormous. Alternatively if the correct guess was 0 and we guessed 0, our cost function should not add any cost for each time this happens. If the guess was right, but we weren’t completely confident (e.g. y = 1, but h(x) = 0.8), this should come with a small cost, and if our guess was wrong but we weren’t completely confident (e.g. y = 1 but h(x) = 0.3), this should come with some significant cost, but not as much as if we were completely wrong.

This behavior is captured by the log function, such that:

cost expressed as log

Again, the cost function J of theta gives us the average cost over all of our training examples.

So here we’ve described how the predictor h(x) and the cost function J of theta differ between regression and classification, but gradient descent still works fine.

A classification predictor can be visualized by drawing the boundary line; i.e., the barrier where the prediction changes from a “yes” (a prediction greater than 0.5) to a “no” (a prediction less than 0.5). With a well-designed system, our cookie data can generate a classification boundary that looks like this:

A graph of a completed machine learning example using the sigmoid function.

Now that’s a machine that knows a thing or two about cookies!

An Introduction to Neural Networks

No discussion of Machine Learning would be complete without at least mentioning neural networks. Not only do neural nets offer an extremely powerful tool to solve very tough problems, but they also offer fascinating hints at the workings of our own brains, and intriguing possibilities for one day creating truly intelligent machines.

Neural networks are well suited to machine learning models where the number of inputs is gigantic. The computational cost of handling such a problem is just too overwhelming for the types of systems we’ve discussed above. As it turns out, however, neural networks can be effectively tuned using techniques that are strikingly similar to gradient descent in principle.

A thorough discussion of neural networks is beyond the scope of this tutorial, but I recommend checking out our previous post on the subject.

Unsupervised Machine Learning

Unsupervised machine learning is typically tasked with finding relationships within data. There are no training examples used in this process. Instead, the system is given a set data and tasked with finding patterns and correlations therein. A good example is identifying close-knit groups of friends in social network data.

The Machine Learning algorithms used to do this are very different from those used for supervised learning, and the topic merits its own post. However, for something to chew on in the meantime, take a look at clustering algorithms such as k-means, and also look into dimensionality reduction systems such as principle component analysis. Our prior post on big data discusses a number of these topics in more detail as well.


We’ve covered much of the basic theory underlying the field of Machine Learning here, but of course, we have only barely scratched the surface.

Keep in mind that to really apply the theories contained in this introduction to real life machine learning examples, a much deeper understanding of the topics discussed herein is necessary. There are many subtleties and pitfalls in ML, and many ways to be lead astray by what appears to be a perfectly well-tuned thinking machine. Almost every part of the basic theory can be played with and altered endlessly, and the results are often fascinating. Many grow into whole new fields of study that are better suited to particular problems.

Clearly, Machine Learning is an incredibly powerful tool. In the coming years, it promises to help solve some of our most pressing problems, as well as open up whole new worlds of opportunity for data science firms. The demand for Machine Learning engineers is only going to continue to grow, offering incredible chances to be a part of something big. I hope you will consider getting in on the action!


This article draws heavily on material taught by Stanford Professor Dr. Andrew Ng in his free and open Machine Learning course. The course covers everything discussed in this article in great depth, and gives tons of practical advice for the ML practitioner. I cannot recommend this course highly enough for those interested in further exploring this fascinating field.

Understanding the Basics

What is Deep Learning?

Deep learning is a machine learning method that relies on artificial neural networks, allowing computer systems to learn by example. In most cases, deep learning algorithms are based on information patterns found in biological nervous systems.

About the author

Nick McCrea, United States
member since February 5, 2014
Nicholas is a professional software engineer with a passion for quality craftsmanship. He loves architecting and writing top-notch code, and is proud of his ability to synthesize and communicate ideas effectively to technical and non-technical folks alike. Nicholas always enjoys a novel challenge. [click to continue...]
Hiring? Meet the Top 10 Freelance Machine Learning Engineers for Hire in March 2019


Great summary! Just a small typo "rpaidly", right after the image of the robot.
Great article, can think of a number of uses in own field!
H. Singer
Thanks for catching this. We've corrected it. Much appreciated.
Julien L
Best toptal blog post ever published so far!
Halfway through this I was wondering whether it was Ng's syllabus summarized. Nice touch!
Vlad Miller
Yeah, looks like its summarize of first four weeks of Andrew Ng lessons.
Sebastián Tramontana
Awesome post!!!
Great post. Here's one more to correct: “feels it’s way” should be “feels its way”, and "voila" could be "voilà" :)
Nick McCrea
Done! Great suggestions. Thanks.
Rob M
Your cookie data is basically the same data Andrew used for assignment 2 on microchips, replotted with a slightly different scale. Good post, but I think you should cite the graph correctly rather than a blanket acknowledgement
The greatest tutorial I've read
Nick McCrea
Awesome! Glad you liked it.
very nice tutorial. what did you use to create the images?
Nick McCrea
The images are screenshots of Octave's graphing functions. Toptal's artist then framed the screenshots in a nice Toptal frame.
Aakash Kag
very usefull information for begineer ...great job sir ..thanx lot
Pedro Mauricio Corvo
Fantastic - thank you for the clear cut explanation
Joey Goertzen
As long as we are correcting errors, here is another one: "Instead of using brute force, an machine learning system..." => "Instead of using brute force, a machine learning system" or "Instead of using brute force, an ML system"
Joey Goertzen
Found another one: "The answers lies in our “measurement of wrongness”" => "The answer lies in our “measurement of wrongness” "
Ramana Murthy
An excellent intro - thanks
Jan Lanik
"With least squares, the penalty for a bad guess goes up exponentially with the difference between the guess and the correct answer" Quadratically, not exponentially :) Otherwise, nice intro :)
Thank you for this article. It maybe an introduction but it is of immense value.
Thank you very much. Very useful post for a beginner.
Andrew Ng content is nicely summarized - Can we intent to add more practical business use cases - in the fields of Banking , Retail and how it can be applied , pl suggest
Sheila Cloudcroft
Thank you!!!!!!!! I had no idea what ML was. This is a great intro. I super appreciate this.
Same thought. Better, I think, to mention that near the top.
I came here on year after you lol ;)
Oleksandr Negodiuk
There is trasformations criteria?
Areeg Samir
I read about the online and offline algorithms but I don't know to which type (supervised or unsupervised) they are related. Thank you it is a good article.
Anil Singh
Awesome post. Thanks http://www.code-sample.com/
This was great--thanks! Thought you might also enjoy this new Machine Learning Tutorial: https://www.praetorian.com/blog/machine-learning-tutorial This post gives a neat example of machine learning on binary data. If you're familiar, the author also released a technical challenge on the topic at https://mlb.praetorian.com
Hello, this is really a very nice tutorial. I have searched on this topic before but I got lots of boring stuffs which I could not understand. Am I allowed to post this on my blog with proper credits,citations and link referal to you ? Thanks. Vicky (techvicky.com)
the classification problem looks exactly like maximum likelihood estimation (the first example is infact a sub-category of max likelihood i.e. ordinary least squares), is there any real difference between mathematical statistics and machine learning? eager to know. was thinking of reading few books on machine learning but looks like a repeat. regards.
Barbara Horne
Nice. I learned a lot from this. Thank You.
Nadeem Shaikh
consise explanation
ankit mistry
Great summary about Machine Learning introduction. Check this another video series on Machine Learning : https://www.youtube.com/playlist?list=PLA83b1JHN4lxUAZC7a3vSs2DU8lnDxez6
Really helpful post
Timothy Mayabi
awsome. You have saved my neck just before exams
Vikas Gupta
Nice tutorial. Cleared some doubts. Thanks.
shivendra pratap singh
I came here after one more year :P
Pranav Makkar
i came here after a day .. :D
Good article. thank you for explaining.
Joseph Kim Dadasonicson
Nice post...helped a lot. Let me share this. <a href="https://www.mnlearn.me">Machine Learning</a>
Prithvi Raj
Are you sure the classification boundary in the case of a sigmoid function can be elliptical shaped as you have shown? I thought a logistic regression's classification boundary was a straight line.
Saygın Burak Sevinç
Some images are not displayed on page. 404 not found.
It Enterprise
Thanks For Posting Important Article on <a href=" www.itenterprise.co.uk "> Machine Learning Examples </a>
ankit mistry
Simple layman introduction of machine learning. https://www.youtube.com/watch?v=__snatmeLM0
ankit mistry
Nice video, Nice explanation.
Illahi Bux
i cam here after 2 years
almalky leader
Question: how you can find the h(6) , θ0=−2, θ1=0.5, Here each row is one training example. Recall that in linear regression, our hypothesis is hθ(x)=θ0+θ1x, and we use m to denote the number of training examples. Row x= 3,1,0,4 \ Row y=2,2,1,3
Miles Alden
Excellent tutorial. I'm learning this stuff from scratch and am wondering how one would go about "training" in your cookie example?
Preksha Das
Hi, This blog is nice and some where informative as well. You have mentioned all the things in a well mannered way which is really good and it seems quite impressive as well. You have given a information in a detailed mannered. I have used one website before for mine work where i had been a great experienced which was great for mine career growth and as well as mine career exposure(http://www.hifreelancer.com). So, i would suggest to people should give it a try once might be they would find great place work there and then looking forward to more work over there. Thank you for the post
Selim Chehimi
Really nice articles, I learned a lot. I post daily about artificial intelligence, you might wanna check it out : http://selimchehimi.com
Nicole M. Blewett
Peter Lusink
Thank you Nick! Great article :) For anyone who is interested in Machine Learning from a more ‘business point of view’ I got myself online certified at AIcompany (aicompany.co) in order to understand how machine learning can be integrated in my line of business. They also have management frameworks that helps getting successful pilot started. For me a focus on best practices and implementation is important. Anyway, understanding the basics of machine learning is (nowadays) a must and a great way to have a more distinctive resume.
Francesca Estrada
Hi Sir! May I ask when did you publish this article? My team and I are currently working on our thesis that is related to machine learning and we would like to cite your work. Thank you very much.
Still browsing ;)
Pradeep K. Pant
Great summary with beautiful images. actually I took NG MI course last summer and was trying to refresh week per week. Meanwhile found your summary. Good work. Thanks!
Excellent post
Christia Buntleitner
Almost asked the same question)))
johny blaze
Thanks For the Valuable Information About the Machine Learning and its Benefits of Using this Technologies on the Present Career Job Oriented Courses. For Any Information about Machine Learning Please Visit http://www.analyticspath.com/machine-learning-training-in-hyderabad .
johny blaze
Very Pleased to receive the information about machine learning and professional courses and the concepts.for further information about machine learning please visit http://www.analyticspath.com/machine-learning-training-in-hyderabad
i came all over here
neel bhatt
Nice post. I have started serires of posts on Machine Learning and Interview Q&A. Have a look here: https://neelbhatt40.wordpress.com/2017/10/29/machine-learning-in-simple-words-azure-machine-learning-part-i/
Dr. Fatih Tekin
The first predictor equation h(x) = 12.00 +0.20 x should be h(x) = 20.00 +0.12 x. The plot below the predictor also suggests this equation. Otherwise the prediction of satisfaction for employees with $60k salary would not be 27 but 24 (12.00+0.20*60 = 24)
I came here after 3 years.
Rahil Sharma
Thanks for introduction to ML in such a low level manner .you made it very easy to understand for a beginner like me.
Rameez Ramzan Ali
I had read numerous posts on ML and AI but on one discussed about pros and cons of ML and AL. If everyone start working on ML then what would be future of ML and AI? (https://www.cubix.co/blog/machine-learning-ai/)
Thank you for this article. Machine learning is gaining lot of importance since it can be used to solve complex problems and also improves user experience.
Prakhar Vyas
I don't know how i landed up here xD
Yuvan Asav
Great article about the machine learning.After reading this article,got to know It's really interesting language.Now a days, everyone is talking about machine learning and big data. Thanks for Sharing the interesting topic. I would like to add one more topic related to it https://www.springpeople.com/blog/python-vs-r-for-machine-learning-which-is-better/
I'm still here after all these years...
Andrea Maria
The process of learning needs specially built algorithms that would teach machines what exactly they have to do. Machine learning has its applications in banking industry, finance industry, healthcare industry, data mining and in robotics.
Bo Ang Naniwalanaman
I came here before ya all were born.
Divya Singh
Good read! Got few basics clarified about ML. For info. on Machine Learning Trends in 2018, check here : <a href = "https://www.springpeople.com/blog/5-hot-ai-machine-learning-trends-for-2018/" > Machine Learning trends for 2018 </A>
Peter Tornyos
I've got another very good article you might also consider to read.. https://www.javaworld.com/article/3224505/application-development/machine-learning-for-java-developers.html
Would you please post the liner regression training data( in your post: wherein company employees have rated their satisfaction) here? Thank you very much!
Raj G
Nice article on Machine learning. You may also get latest info in www.DataScienceTutor.com
Tim Maes
Time for me to come now
shivendra pratap singh
time flies so fast. Now I hope every year someone new will carry this chain forward.
Heraclio Munoz
To best understand how machine learning works, let’s use the example of how streaming services generate movie recommendations for their subscribers. In order to recommend the types of movies individual subscribers would be interested in watching, machine learning algorithms analyse the viewing preferences of that individual and compare them with the data of other users with similar tastes in movies. The algorithm then learns which movies to recommend based on its experience of doing so. The ability to learn from their own experience is what has made machine learning algorithms so attractive to businesses and enterprises in operating in a plethora of different occupational spaces, from finance and healthcare to farming and cyber security. However, while machine learning algorithms are more than capable of improving their performance overtime, they do still need the occasional bit of human oversight. If their performance results are unsatisfactory or anomalous, humans will still need to step in and make adjustments or change certain parameters. Artificial Intelligence – Part 2 – Deep Learning Vs. Machine Learning: Understanding the Difference: https://www.lanner-america.com/blog/artificial-intelligence-part-2-deep-learning-vs-machine-learning-understanding-difference/
Super insightful article. For anyone looking to practice machine learning, here's a list of the best 50 free datasets on the web: https://gengo.ai/datasets/the-50-best-free-datasets-for-machine-learning/
Presmelito Rapatan Villanueva
pres was here
Gilda Dahis
Thank you for your enlightening introduction to this important topic. Your explanations and examples were very helpful to me.
cetpa seo
hello i started coaching for machine learning a week ago and was soo confused about what i am doing , After going through this article it help me alot to build concept about Machine Learning. it would be helpful if you keep on updating more article on machine learning. thank you
comments powered by Disqus
Free email updates
Get the latest content first.
No spam. Just great articles & insights.
Free email updates
Get the latest content first.
Thank you for subscribing!
Check your inbox to confirm subscription. You'll start receiving posts after you confirm.
Trending articles
Relevant Technologies
About the author
Nick McCrea
Software Developer
Nicholas is a professional software engineer with a passion for quality craftsmanship. He loves architecting and writing top-notch code, and is proud of his ability to synthesize and communicate ideas effectively to technical and non-technical folks alike. Nicholas always enjoys a novel challenge.