CAPITAL CORP. SYDNEY

73 Ocean Street, New South Wales 2000, SYDNEY

Contact Person: Callum S Ansell
E: callum.aus@capital.com
P: (02) 8252 5319

WILD KEY CAPITAL

22 Guild Street, NW8 2UP,
LONDON

Contact Person: Matilda O Dunn
E: matilda.uk@capital.com
P: 070 8652 7276

LECHMERE CAPITAL

Genslerstraße 9, Berlin Schöneberg 10829, BERLIN

Contact Person: Thorsten S Kohl
E: thorsten.bl@capital.com
P: 030 62 91 92

Neural Network & Deep Learning

Neural Network & Deep Learning

Introduction to Neural Network and Deep Learning

1.       Introduction to Neural Networks

2.       NN History

3.       Why data-driven?

4.       Handson Python demo of K-NN classifier on MNIST dataset

5.       The choice of distance is a Hyperparameter

6.       K-NN never used on images

7.       Parametric Approach

8.       Linear Algebra_Scalars, vectors, matrices, tensors

9.       Algebra_Linear operations on vectors and matrices

 

Mathematics for Deep Learning : Linear Algebra part 1

1.       Scalars, Vectors, Matrices, And Tensors

2.       Linear operations on Vectors and Matrices

3.       Vector Properties: Vector norms, some special vectors and matrices

4.       Hands-on_python demo: vector and matrix properties

Hands On Lab

1.       LinearAlgebra.ipynb

2.       Linear Algebra.pptx

3.       Matrices

4.       Line Concept

5.       DLCP Project 1 Brief.pdf

6.       KNN_SVHN.ipynb

7.       Link to the dataset

Math for Deep Learning : Functions and Convex optimization

1.       GPU Cloud Labs : Google Colabs

1.       Working with Google Colabs Part1: Getting Started

2.       Working with Google Colabs Part2: Mounting Data

2.       Functions and derivatives part 1

3.       Functions and derivatives part 2

4.       Optimizing a continuous function part 1

5.       Optimizing a continuous function part 2

6.       Hands-on Python demo – Gradient descent

7.       GradientDescent2.ipynb

Introduction to Loss Functions

1.       Components of Supervised Machine learning

2.       Components of Supervised ML: Model, Parameters, and Hyper parameters

3.       Components of Supervised ML: Loss functions

4.       Examples of Loss functions: MSE vs MAE Loss for Regression

5.       Hands-on python demo: MSE loss function

6.       SimpleLossAndGradientDescent.ipynb

7.       Examples of loss functions for classification: MSE vs Cross Entropy loss

8.       Hands-on Python demo: MSE vs CE Loss functions

9.       SimpleLossAndGradientDescentV2.ipynb

10.   Regularization

Neural Networks deconstructed for Supervised Learning (Classification)

1.       Introduction to Neural Networks

2.       Activation functions

3.       Feed forward neural network

4.       Back propagation and Gradient descent

5.       Learning Rate setting and tuning

6.       Hands-on Python Demo: Building a Neural Network from Scratch

7.       NN_MNIST_Scratch_v1.ipynb

8.       Hands-on Python Demo: Visualizing binary decision boundaries

9.       NN from scratch with boundary visualized v1.ipynb

 

Building Blocks of Neural Networks

1.       Feed forward

2.       Back Propagation

3.       Fully Connected Layer – Forward pass

4.       Fully Connected Layer – Backward pass

5.       Activation Functions

6.       Activation functions – In practice

7.       Softmax

8.       Cross-Entropy Loss

9.       Hands-on-Python-demo

1.       MNIST walk-through of building blocks of NN

2.       MNIST Python Neural Network_Final.ipynb

10.   Hands On Lab

1.       DLCP Project 1 Brief.pdf

2.       SVHN Python Neural Network_Milestone2_Questions.ipynb

3.       SVHN Python Neural Network_Solution_notebook.ipynb

 

TensorFlow, Keras, and Tensorboard

1.       Introduction to Tensorflow

2.       Computational Graph

3.       Hands-on in TensorFlow: Linear regression on Boston Housing prices

4.       Introduction to Keras

5.       Build a Deep Neural Network in Keras: MNIST Dataset

6.       Using Tensorboard

7.       Codes used in the module

8.       Tensorflow_Hello_World.ipynb

9.       Boston_Housing_Prices.ipynb

10.   Classification_MNIST_DNN_Keras.ipynb

11.   Boston_Housing_Prices_KERAS.ipynb

Babysitting the Neural Network

1.       Introduction to babysitting the learning process

2.       Data Preprocessing

3.       Data Augmentation

4.       Weight initialization

5.       Regularization – Batch Normalization

6.       Regularization – Dropout

7.       Hands on python demo – Babysitting the neural network and Hyperparameter optimization

8.       Babysitting-MNIST Python Neural Network-FINAL.ipynb

9.       Visualizations

Project Work

1.       DLCP Project 1 Brief.pdf

2.       Babysitting-SVHN Python Neural Network_Questions.ipynb

3.       Babysitting_SVHN Keras Neural Network_Questions.ipynb

4.       Babysitting_SVHN Keras Neural Network_Solution_notebook.ipynb

5.       Babysitting_SVHN Python Neural Network_Solution_notebook.ipynb