Machine learning

Machine learning is a branch of computer science aimed at enabling computers to learn new behavior based on empirical data. The goal is to design algorithms that allow the computer to display behavior learned from past experience, rather than human instruction. Machine learning is essential to the development of artificial intelligence, but it is also applicable to many everyday computing tasks. Common examples of programs that employ machine learning include email spam filters, optical character recognition, and news feeds on social networking sites that alter their displays based on previous user activity and preferences.

90558380-100594.jpg

Overview

One of the earliest attempts to enable machine learning was the perceptron algorithm, developed by Frank Rosenblatt in 1957. The algorithm was intended to teach pattern recognition and was based on the structure of a neural network, which is a computing model designed to imitate an animal’s central nervous system. While the perceptron model showed early promise, Marvin Minsky and Seymour Papert demonstrated in their 1969 book Perceptrons: An Introduction to Computational Geometry that it had significant limitations, as there were certain classes of problems the model was unable to learn. Consequently, researchers did not pursue this area of study for some time.

In the 1980s, after other avenues for developing artificial intelligence had resulted in only limited success, scientists began to revisit the perceptron model. Multilayer perceptrons, or neural networks composed of multiple layered computational units, proved to have the processing power to express problems that Rosenblatt’s single-layer, or linear, perceptrons could not. Around the same time, John Ross Quinlan introduced decision-tree algorithms, which use predictive models to determine a variable’s value based on available data.

Since then, numerous machine-learning algorithms have been developed. Among those most commonly used were support vector machines (SVMs) and naive Bayes classifiers. SVMs, introduced by Vladimir N. Vapnik and Corinna Cortes in 1995 and based on an algorithm previously created by Vapnik, were used to recognize patterns in data and classify the various data points. Naive Bayes classifiers are applications of Bayes’s theorem, named for the eighteenth-century mathematician and reverend Thomas Bayes, which deals with conditional probabilities. This algorithm was used in one of the earliest e-mail spam filters, iFile, released by Jason Rennie in 1996. Many e-mail clients employed Bayesian spam filtering, which worked by determining the probability that an e-mail containing certain keywords is spam.

Machine-learning algorithms can be divided into categories based on how they train the machine. These categories include supervised learning, in which the machine learns from inputs that are mapped to desired outputs; unsupervised learning, in which the machine analyzes input without knowledge of the desired output; semi-supervised learning, in which some of the input is paired with a desired output and some is not; transduction, in which the machine tries to predict new outputs based on training with previous inputs and outputs; reinforcement learning, in which the machine must form a policy on how to act based on observing how certain actions affect its environment; and learning to learn, which teaches inductive bias based on previous experience. SVMs, multilayer perceptrons, decision trees, and naive Bayes classifiers all fall into the category of supervised learning.

It is important to distinguish between machine learning and data mining. Although the two concepts are related and use similar, often overlapping, methods, data mining is focused on discovering information about given data, while machine learning focuses more on learning from the given data in order to make predictions about other data in the future. Many consider data mining to be a subset of machine learning.

The field of machine learning experienced several breakthroughs in the twenty-first century, particularly surrounding the emergence of new AI programs that exhibited pronounced advancements when compared to previous iterations. One such example was ChatGPT, an AI chatbot developed by AI research group OpenAI capable of using a vast wealth of information to generate complex responses to user prompts. The program was incredibly popular following its release in 2022 and found practical uses in business, research, coding, and academia, among many other fields. However, the program also proved controversial in many ways, including its ability to trivialize certain academic assignments. ChatGPT utilized the systems of supervised learning and reinforcement learning to operate.

Bibliography

Abu-Mostafa, Yaser S. “Machines That Think for Themselves.” Scientific American, July 2012, pp. 78–81.

Brodley, Carla E. “Challenges and Opportunities in Applied Machine Learning.” AI Magazine, vol. 33, no. 1, 2012, pp. 11–24.

Domingos, Pedro. “A Few Useful Things to Know about Machine Learning.” Communications of the ACM, vol. 55, no. 10, 2012, pp. 78–87.

Heaven, Douglas. “Higher State of Mind.” New Scientist, 10 Aug. 2013, pp. 32–35.

Marsland, Stephen. Machine Learning: An Algorithmic Perspective. Taylor, 2009.

Mitchell, Tom M. The Discipline of Machine Learning. Carnegie Mellon U, 2006.

Mitchell, Tom M. Machine Learning. McGraw, 1997.

Piore, Adam. “Mind in the Machine.” Discover, June 2013, pp. 52–59.

Ramponi, Marco. “How ChatGPT Actually Works.” AssemblyAI, 23 Dec. 2022, www.assemblyai.com/blog/how-chatgpt-actually-works/. Accessed 27 Aug. 2024.

"The Revolution of Deep Learning in 2012: A Paradigm Shift in Artificial Intelligence." Medium, 22 Sept. 2023, medium.com/@kaveh.kamali/the-revolution-of-deep-learning-in-2012-a-paradigm-shift-in-artificial-intelligence-d4fdbfa87a42. Accessed 27 Aug. 2024.

"What Is Machine Learning (ML)?" IBM, www.ibm.com/topics/machine-learning. Accessed 27 Aug. 2024.