Additional Resources

Part 6

Throughout these tutorials, we learnt what a neural network was and how it processed data using perceptrons and layers. Then we continued by exploring how these networks actually learn through the processes of gradient descent and backpropagation. We looked at the different types of neural networks, from CNNs A Convolutional Neural Network (CNN) is a type of neural network designed to process data with a grid-like structure, such as images. Instead of looking at the entire image at once, it uses small filters (kernels) that slide over the input to detect local patterns like edges, textures, or shapes, and then combines these to recognise more complex features. and RNNsA Recurrent Neural Network (RNN) is a type of neural network designed to handle sequential data by carrying information from one step of the sequence to the next through its hidden state. This memory-like ability makes it useful for tasks such as speech recognition, language modeling, and time-series prediction., to TransformersA Transformer is a neural network architecture that processes entire sequences in parallel using a mechanism called self-attention, which allows it to weigh the importance of different parts of the input relative to each other. This design makes it especially powerful for tasks like translation, text generation, and large-scale language modeling. and GANsA Generative Adversarial Network (GAN) consists of two competing neural networks: a generator that creates synthetic data and a discriminator that tries to distinguish between real and fake data. Through this adversarial training, GANs learn to produce highly realistic outputs, such as images, music, or even video.. Finally, we finished off by having a look at what implementing a neural network in real life involves, and how neural networks are used in a variety of fields to solve huge problems today. I hope I've taught you something, and possibly even encouraged you to dig deeper to elements of neural networks that you are curious about. I have compiled some good free online resources should you wish to continue this journey:

▬▬▬▬▬▬


"Vanilla" Neural Networks (FNNs):

▬▬▬▬▬▬


Transformers and LLMs:

▬▬▬▬▬▬


CNNs:



Glossary and Frequently Used terms:

Artificial Intelligence (AI)
Machines designed to perform tasks that usually require human intelligence, such as understanding speech, recognising images, or making decisions.
Machine Learning (ML)
A subset of AI where machines learn from data to make predictions or decisions without being explicitly programmed.
Deep Learning
A specialised area within ML that uses multi-layered neural networks to model complex patterns in large datasets. It is called "deep" because of the many layers of processing involved.
Neural Network
A computational model inspired by the human brain, consisting of layers of nodes (neurons) that process input data to produce an output.
Neuron
The basic unit in a neural network that receives input, processes it, and passes on its output to the next layer.
Layer
A collection of neurons in a neural network. Common types include:
  • Input Layer: Receives raw data.
  • Hidden Layers: Process data through weighted connections.
  • Output Layer: Produces the final prediction or classification.
  • Activation Function
    A mathematical function applied to a neuron's output, introducing non-linearity to help the network learn complex patterns.
    Training
    The process of teaching a neural network by feeding it data and adjusting its weights to minimise errors in predictions.
    Backpropagation
    A method used during training where the network adjusts its weights by calculating the gradient of the error and propagating it backwards through the network.
    Epoch
    One complete pass through the entire training dataset.
    Overfitting
    When a model learns the details and noise in the training data to the extent that it negatively impacts its performance on new data.
    Underfitting
    When a model is too simple to capture underlying patterns in the data, leading to poor performance on both training and new data.
    Generalisation
    The ability of a model to perform well on new, unseen data.
    Variance
    The model's sensitivity to small fluctuations in the training dataset. High variance can lead to overfitting.
    Cost/Loss Function
    A mathematical function that measures the difference between the model's predictions and the actual outcomes. The goal is to minimise this loss.
    Optimiser
    An algorithm used to adjust the weights of the neural network to minimise the cost/loss function.
    Convolutional Neural Network (CNN)
    A type of neural network particularly effective for processing structured grid data, such as images.
    Recurrent Neural Network (RNN)
    A type of neural network designed for sequential data, such as time series or natural language.
    Transfer Learning
    Reusing a pre-trained model on a new but related problem, saving time and computational resources.