Machine learning is a rapidly growing field that has gained immense popularity in recent years. It is a subset of artificial intelligence that focuses on providing computers with the ability to learn and improve from data without being explicitly programmed. With its ability to identify patterns and make decisions based on the data it is trained on, machine learning has found applications in various domains such as healthcare, finance, engineering, medicine, and marketing. Here we provide an overview of machine learning, its types, applications, and techniques, and discuss the challenges it faces. We will also look at some real-world examples of machine learning in action and explore the future of this exciting field.
An entity that performs behaviours that a person might reasonably call intelligent if a human were to do something similar. A broader simplification is that if we look at something artificial and it does things that are engaging and useful and non-trivial, then we might call it intelligent. The important thing to note is that AI is not magic because it can be explained.
A means by which to create behaviour by taking in data, forming a model, and then executing the model The model reference here is a simplification of some complex phenomenon. For example, a model car is just a smaller, simple version of a real car. Just like we can make a smaller, simpler version of a car, we can also make a smaller, simpler version of human language. we use the term large language Models because these are, well large, from the perspective of how much memory is required to use them. The largest models in production such as GPT-3 and GPT-4 are large enough that it requires massive super-computers running in data center servers to create and run
There are many ways to learn a model from data. The Neural Network is one such way. The technique is roughly based on how the human brain is made up of a network of interconnected brain cells called neurons that pass electrical signals back and forth, somehow allowing us to do all the things we do. The basic concept of the neural network was invented in the 1940s and the basic concepts of how to train them as were invented in the 1980s. Neural networks are very inefficient, and it wasn’t until around 2017 that computer hardware was good enough to use them at large scale.
We can look at text written by humans and wonder whether a circuit could produce a sequence of words that looks a lot like the sequences of words that humans tend to produce. We are trying to design an algorithm that guesses an output word, given a bunch of input words.
For example: " The universe is extremely ---- " seems like it should fill in the blank with "large" but not "voodoo".
We tend to talk about language models in terms of probability. We would expect a good language model to produce a higher probability of the word "large" than the word "Voodoo" in the simple example above.
ChatGPT together with GPT-3 and GPT-4 are all the rage where ever you look. GPT is a particular branding of a type of large language model developed by a company called OpenAI. GPT stands for Generative Pre-trained Transformer.
Let's break this down:
Watch The Video