Blog
Artificial intelligence, machine learning, deep learning and natural language processing – when the conversation turns to artificial intelligence and the ubiquitous buzzwords begin to fly, non-developers soon switch off. These terms are often used too broadly, synonymously, or simply incorrectly. To shed a little light on this matter, and to help you deal with AI applications successfully, we thought we would clear up any terminological confusion.
And where better to start than at the beginning? So, first things first:
Artificial intelligence (AI for short) is a general term covering all fields of research investigating the application of human intelligence by machines.
Often AI is further subdivided into “strong” and “weak” AI. Weak AI, which is in use today, can effectively simulate individual human capabilities, while strong AI, which is still a long way from reality, can think, reason and act as independently as a human being.
Machine learning (also known as ML), a term often used synonymously with artificial intelligence, is actually a sub-field of AI.
ML concerns the capacity to learn autonomously. So it is not just about imitating processes (or intelligence), but also about using experience to generate knowledge.
In the case of machine learning, this means learning statistical patterns from representative historical data. Today, this enables processes which previously had to be carried out manually to be fully or partly automated and thus optimised.
For example, all a modern machine learning-based image recognition system needs to produce reliable results is examples of correctly tagged images. The machine learning algorithm uses this representative data to learn how to identify specific objects. As it would be too complex to write general if-then rules precisely defining an object’s appearance at pixel level, a rules-based expert system would probably never achieve the same performance.
Unlike machine learning algorithms, rules-based systems are also non-transferable. A system that classifies German texts using rules developed manually could probably not be used for Chinese texts. However, if data for German and Chinese are available, there is a good chance that, after language-specific pre-processing, the same machine learning algorithm could learn how to classify both German and Chinese texts as required.
Consequently, for a rules-based expert system faced with a complex problem, the initial development process is long and costly. And because the resulting systems are as complex as the original problem, maintenance and subsequent development is also laborious. By contrast, a machine learning-based system typically has significantly less code because the logic does not need to be explicitly stated, bringing advantages in terms of both cost and performance.
In situations where future decisions can be taken on the basis of knowledge drawn from historical data, machine learning is an ideal tool for full or partial automation of processes – provided that data relating to the problem at hand is already available or can be easily obtained.
Deep learning (also known as DL) is in turn a sub-domain of machine learning. DL systems take the knowledge learned by ML systems further: using large volumes of historical data, they act and evolve autonomously, i.e. with no need for existing knowledge to be programmed in beforehand.
They can do this because of the way they operate, which is inspired by the neuronal networks in the human brain. Thanks to their complex (and therefore deeper) structure, systems of this kind use relevant training data to learn basic patterns and infer more complex patterns independently. Deep learning models can become so complex that it is not possible to control or track exactly what they learn. And while DL too is limited to the problems dealt with in the data it is presented with, it can, under the right conditions, produce significantly better results than simpler ML algorithms.
Deep learning requires enormous volumes of data to search for these patterns. Typical applications for DL today include facial, object, and speech recognition.