Machine learning vs AI
These two terms are often used interchangeably, but they are fundamentally different technologies
We’re in the midst of a fascinating era of innovation. The standard of certain digital technologies is accelerating to the extent that what was limited to the realms of science fiction just a few decades ago has now become a reality.
With technologies advancing, however, so too is the level of complexity involved in understanding and explaining how they work, and even what they are.
This is especially true for certain terms such as Artificial intelligence (AI) and machine learning (ML). They may seem similar, falling under the same broader category, but are significantly different in a number of ways.
While they’re often used interchangeably, particularly in corporate promotion and advertising, they’re not actually the same thing. The confusion originates in the fact that AI is an umbrella term that’s applied broadly to anything that’s deemed ‘smart’ in some way. Smart speakers such as Alexa and Siri, for example, are popularly considered to be ‘AI’, as are virtual assistants embedded onto websites and services.
These are among the most commonly used interfaces when it comes to AI, but the technology is also applied in several business contexts, such as in automating the back office for many organisations. AI is also embedded in other scenarios including intelligent load balancing, for example, or some network security systems.
Machine learning, meanwhile, is a subset of AI. This means that while ML can be described as a form of AI, AI doesn't include machine learning by default. ML itself also incorporates various subdivisions such as reinforcement learning and deep learning.
What's the difference between ML and AI?
The history of AI is a long one. For thousands of years, humans have dreamt of machines that could 'come to life', behaving and thinking as humans do. There was a time when early computers, due to their 'logical' nature, were also considered a type of artificial intelligence.
In its current manifestation, however, the idea of AI can trace its history to British computer scientist and World War II codebreaker Alan Turing. He proposed a test, which he called the imitation game but is more commonly now known as the Turing Test, where one individual converses with two others, one of which is a machine, through a text-only channel. If the interrogator is unable to tell the difference between the machine and the person, the machine is considered to have "passed" the test.
This basic concept is referred to as 'general AI' and is generally considered to be something that researchers have yet to fully achieve.
However, 'narrow' or 'applied' AI has been far more successful at creating working models. Rather than attempt to create a machine that can do everything, this field attempts to create a system that can perform a single task as well as, if not better than, a human.
It's within this narrow AI discipline that the idea of machine learning first emerged, as early as the middle of the twentieth century. First defined by AI pioneer Arthur Samuel in a 1959 academic paper, ML represents "the ability to learn without being explicitly programmed".
Uses and applications
Interest in ML has waxed and waned over the years, but with data becoming an increasingly important part of business strategy, it's fallen back into favour as organisations seek ways to analyse and make use of the vast quantities of information they collect on an almost constant basis.
When this data is put into a machine learning program, the software not only analyses it but learns something new with each new dataset, becoming a growing source of intelligence. This means the insights that can be learnt from data sources become more advanced and more informative, helping companies develop their business in line with customer expectations.
One application of ML is in a recommendation engine, like Facebook's newsfeed algorithm, or Amazon's product recommendation feature. ML can analyse how many people are liking, commenting on or sharing posts or what other people are buying that have similar interests. It will then show the post to others the system thinks will like it.
ML is also particularly useful for image recognition, using humans to identify what's in a picture as a kind of programming and then using this to autonomously identify what's in a picture. For example, machine learning can identify the distribution of the pixels used in a picture, working out what the subject is.
Enterprises are now turning to ML to drive predictive analytics, as big data analysis becomes increasingly widespread. The association with statistics, data mining and predictive analysis have become dominant enough for some to argue that machine learning is a separate field from AI.
The reason for this is that AI technology, such as natural language processing or automated reasoning, can be done without having the capability for machine learning. It is not always necessary for ML systems to have other features of AI.
There are hundreds of use cases for AI, and more are becoming apparent as companies adopt artificial intelligence to tackle business challenges.
One of the most common uses of AI is for automation in cyber security. For example, AI algorithms can be programmed to detect threats that may be difficult for a human to spot, such as subtle changes in user behaviour or an unexplained increase in the amount of data being transferred to and from a particular node (such as a computer or sensor). In the home, assistants like Google Home or Alexa can help automate lighting, heating and interactions with businesses through chatbots.
There are well-founded fears that AI will replace human job roles, such as data input, at a faster rate than the job market will be able to adapt to. Author and venture capitalist Kai-Fu Lee, who has worked at both Apple and Google and earned a PhD from Carnegie Mellon for the development of an advanced speech recognition AI, warned in 2019 that "many jobs that seem a little bit complex, a chef, a waiter, a lot of things, will become automated."
"We will have automated stores, automated restaurants and all together, in 15 years, that's going to displace about 40% of jobs in the world."
Confusing AI and ML
To make matters more confusing when it comes to naming and identifying these terms, there are a number of other terms thrown into the hat. These include artificial neural networks, for instance, which process information in a way that mimics neurons and synapses in the human mind. This technology can be used for machine learning; although not all neural networks are AI or ML, and not all ML programmes use underlying neural networks.
As this is a developing field, terms are popping in and out of existence all the time and the barriers between the different areas of AI are still quite permeable. As the technology becomes more widespread and more mature, these definitions will likely also become more concrete and well known. On the other hand, if we develop generalised AI, all these definitions may suddenly cease to be relevant.
The ultimate law enforcement agency guide to going mobile
Best practices for implementing a mobile device programFree download
The business value of Red Hat OpenShift
Platform cost savings, ROI, and the challenges and opportunities of Red Hat OpenShiftFree download
Managing security and risk across the IT supply chain: A practical approach
Best practices for IT supply chain securityFree download
Digital remote monitoring and dispatch services’ impact on edge computing and data centres
Seven trends redefining remote monitoring and field service dispatch service requirementsFree download