Deep learning, machine learning, and pattern recognition are highly relevant topics commonly used in the field of robotics with artificial intelligence. Despite the overlapping similarities, these concepts are not identical. In this article, we will be discussing some of the differences of the three concepts and their applications.
Figure 1: An algorithm to detect the character "3" using sub-blocks
Pattern recognition is the oldest form of learning and has become a relatively obsolete term. On the other hand, deep learning is a new and popular topic in the field of artificial intelligence. Machine learning, unlike the other two terms, is a fundamental form of learning and is one of the hottest areas in many start-ups and research labs. The Google Trends image below that shows the recent increase in interest for deep learning. Additionally, from the image, we can conclude that:
Figure 2: The Google search index of the three concepts since 2004 (Picture source: Google Trends)
The term pattern recognition became popular between the 1970s and the 1980s. It focuses on how to make computer programs perform intelligent and human-like tasks, such as an object from an image. Initially, people were not very interested in knowing how machines can achieve this intelligence, as long as it works correctly. However, technologies such as filters, boundary detection, and morphological processing have shown to be effective when applied to an image detection algorithm. Researchers in the pattern recognition community showed an increasing interest in this topic, spawning the field of optical character recognition.
It is appropriate to say that pattern recognition was the most innovative and "intelligent" signal processing of the 1970s, the 1980s, and even the early 1990s. Concepts such as decision tree, heuristic method, and quadratic discriminatory analysis were all introduced during this period. Pattern recognition slowly shifted from being a topic in electrical engineering to a topic of interest in computer science. One of the most famous books in pattern recognition, Pattern Classification by Duda and Hart, was released in 1973. Despite being published more than four decades ago, it is still a good introductory textbook for beginners seeking to know more about the pattern recognition field.
In the early 1990s, many realized that there was a more effective way to create pattern recognition algorithms, particularly replacing researchers with probability and statistics. This paradigm led to the creation of machine learning. The goal of machine learning is to give a computer a collection of data and let the computer make its own conclusion with minimal human intervention. Specifically, it implies that the computer (or the machine) collects statistics from data, and then generating a probabilistic model to determine the most possible outcome. When designed correctly, a machine learning algorithm always performs better than a person would because it is immune to cognitive bias and fatigue.
Figure 3: Typical machine learning process (Picture source: Dr. Natalia Konstantinova's blog).
In the middle of the 21st century, machine learning has emerged as an important research topic in computer science. Scientists have begun applying the concept broadly, creating new businesses using this technology. Machine learning has been used in robotics, genetic analysis, and in predictions for the financial market. Moreover, machine learning's combination with the graph theory created a new topic of research – the graph model.
Machine learning has become a basic skill for many people, but it has also caused a lot of confusion especially to people new to this field. We have seen a wide variety of methods and schools of thoughts in the machine-learning field, all having its own benefits.
Deep learning is currently a hot topic of research, specifically Convolutional Neural Network (or ConvNet), which has been used in large-scale graphic recognition.
Figure 4: ConvNet framework (Picture source: Torch's textbook)
In deep learning, there is minimal human intervention and bias because the parameters in the modes are learned from statistics. However, deep learning is only possible with an ample amount of statistics (big data) and strong arithmetic capabilities (graphic processor or GPU) to optimize the mode. Because convolution computation has been widely applied in computer vision, it is the natural choice for the mode of deep learning.
To understand deep learning, you should have a basic knowledge of linear algebra and programming. If you are not familiar with these topics, we strongly recommend Andrej Karpathy's blog "Hacker's Guide to Neural Networks."
Despite the benefits of deep learning, there are still many unsolved questions in its application. There are no existing theories concerning the validity of deep learning, nor textbooks on specific guidelines of deep learning. There have also been valid concerns for the possibility of artificial intelligence taking over jobs through deep learning. However, successful implementation of deep learning and artificial intelligence still requires plenty of human intervention. A high-quality product requires great vision, expertise of the field, market development, and most of all, the creativity of human beings.
The three popular terms relevant to artificial intelligence – deep learning, machine learning and pattern recognition – are highly correlated but are also unique are used in different applications. Pattern recognition was the first concept to be introduced in image processing, and eventually evolved to machine learning. To maximize the scope of application for machine learning, researchers actively searched for a method to automate machine learning, creating the field of deep learning. Deep learning is a relatively nascent field; there remains a lot of research to explore the full potential of deep learning.
2,623 posts | 723 followersFollow
Alibaba Clouder - May 18, 2021
Alibaba Clouder - April 1, 2021
Alibaba Clouder - January 28, 2021
Alibaba Clouder - January 24, 2017
Alibaba F(x) Team - February 26, 2021
dehong - July 8, 2020
2,623 posts | 723 followersFollow
An end-to-end platform that provides various machine learning algorithms to meet your data mining and analysis requirements.Learn More
This technology can be used to predict the spread of COVID-19 and help decision makers evaluate the impact of various prevention and control measures on the development of the epidemic.Learn More
A high-quality personalized recommendation service for your applications.Learn More
This solution provides you with Artificial Intelligence services and allows you to build AI-powered, human-like, conversational, multilingual chatbots over omnichannel to quickly respond to your customers 24/7.Learn More
More Posts by Alibaba Clouder
Raja_KT February 9, 2019 at 3:49 am
Good demarcation of ML and DL. HPCs , in majority of the cases , demands a lot of compute.