Garth Wales
  • Introduction
  • META
  • To be added elsewhere
  • University
    • 400 Project Research
    • Aims
    • Presentation
    • Report
    • Work Diary
  • Deep Declarative Networks
    • Potential avenues
    • Work Diary
    • Deep Declarative Networks
    • Image Segmentation
  • shortcuts / syntax / setup
    • Jupyter
    • SSH
    • Python
    • CUDA
    • Screen
    • Markdown Syntax
    • Git
    • Dendron
    • Visual Studio Code
    • LaTeX
    • Windows
    • Bash
    • Vim
  • Machine Learning
    • Markov decision process
    • Reinforcement learning
    • Deep Q Network
    • Modern hopfield network
    • Object Detection
    • Face detection
    • Activation Functions
    • Gym
    • RL frameworks
    • Hyperparameters
    • Interpretability
  • Reinforcement learning
    • Memory
    • Exploration techniques
    • Agents
  • garth.wales
    • Website redesign
  • Computer Graphics and Vision
    • Image Mosaicing
    • Stereo Vision
    • Rendering
    • Exam papers
  • Neural Networks
    • Introduction
    • Learning
    • Learning theory
    • Stochastic Neural Networks
    • Interpretability
    • Word Respresentation
    • Language modelling with recurrent networks
    • Sequence-to-sequence networks
    • Exam
Powered by GitBook
On this page
  • Symbolic vs. connectionist models
  • Neural network
  • Advantages
  • Architecture
  • Representation
  • Processing
  • Learning
  • Disadvantages
  • History

Was this helpful?

  1. Neural Networks

Introduction

Symbolic vs. connectionist models

Computation could be thought of as either

  • Continuous sequence of operations (Turing machine! or modern CPU)

    • Computer

      • Memory - static, addressable places

      • Processing - CPU with certain instructions to manipulate following a sequence of rules

      • Learning - Adding to / changing stored memory

    • Fast at doing exactly what you tell it, but needs explicit instructions!

  • Connected graph of events that all feed into feature events (Neural network!)

    • Brain

      • Memory - patterns of activations over multiple neurons

      • Processing - associating patterns, settling to stable states, no 'rules'

      • Learning - modification of weights and biases

    • Map anything to anything

Neural network

Advantages

  • Neural plausibility

  • Parallel tasks inherent (brain is slow, but does much more complex things in parallel compared to CPU)

  • Learn by example! Not a hard coded sequence

  • Content addressable memory (Memory distributed accross neurons)

    • A symbolic system has to learn certain steps

  • Automatic generalisation (learning more general patterns)

  • Robust ('better' than human in many different ways, but worse in others)

    • But can sometimes learn the wrong things... i.e. every picture of a car is on a road so it learns the road is a car..

Architecture

  • Layers, number of units, starting weights

  • Not learnable (unless you use a neural net / genetic algorithm to develop neural nets)

Representation

  • Many layers, many connections per layer

    • What does one neuron do?

    • How do they

Processing

  • Each layer is a weight matrix, each layer has bias vector too

  • Can process multiple matricies at once (batching)

Learning

  • Backpropagation

    • Given an input, outputs something.

    • If the output is wrong, go backwards through the network in order to reduce the error of the output

    • Each neuron should take it's share of the blame

    • Doesn't seem biologically plausible

Disadvantages

  • Distributed representation (Good for GPU training..)

  • Uniqueness (Training from different initial weights will be slightly different..)

  • Modification (How to adjust it to do something new?

  • Interpretation and explanation (Very hard to tell what it does!)

History

a brief recap

  • 1969: Minsky & Papert, 'Perceptron'

  • 1974: Werbos, backpropagation of error

  • 1979: Miyake & Ito, neocognitron (precursor to CNN)

  • 1986: Rumelhart, Hinton & Williams, backpropogation popularised

  • 1990: Elman, simple recurrent neural network

  • 1995: Cortes & Vapnik: Support Vector Machines

  • 1997: Hochreiter & Schmidhuber, Long Short-Term Memory (LSTM) networks

  • 1998: LeCun, convolutional neural network

  • 2012: Krizhevsky, AlexNet

  • 2014: Goodfellow.etal, Generative Adversarial Nets (GAN)

  • 2015: Google Deep Mind’s AlphaGo

  • 2017: Google researchers create Transformer Networks

  • 2018: Generative Pre-trained Transformer (GPT) 2

and more...

PreviousExam papersNextLearning

Last updated 3 years ago

Was this helpful?