Eğitim İçeriği

Part 1 – Deep Learning and DNN Concepts


Introduction to AI, Machine Learning & Deep Learning

  • History, basic concepts, and common applications of artificial intelligence, moving beyond the common fantasies surrounding this field.

  • Collective Intelligence: aggregating knowledge shared by many virtual agents.

  • Genetic algorithms: evolving a population of virtual agents through selection.

  • Usual Learning Machine: definition.

  • Types of tasks: supervised learning, unsupervised learning, reinforcement learning.

  • Types of actions: classification, regression, clustering, density estimation, dimensionality reduction.

  • Examples of Machine Learning algorithms: Linear regression, Naive Bayes, Random Tree.

  • Machine learning VS Deep Learning: problems where Machine Learning remains the state of the art (Random Forests & XGBoost).

Basic Concepts of a Neural Network (Application: multi-layer perceptron)

  • Reminder of mathematical foundations.

  • Definition of a neural network: classical architecture, activation functions.

  • Weighting of previous activations, network depth.

  • Definition of neural network learning: cost functions, back-propagation, stochastic gradient descent, maximum likelihood.

  • Modeling a neural network: modeling input and output data according to the problem type (regression, classification, etc.). The curse of dimensionality.

  • Distinction between multi-feature data and signals. Choosing a cost function based on the data.

  • Approximating a function with a neural network: presentation and examples.

  • Approximating a distribution with a neural network: presentation and examples.

  • Data Augmentation: how to balance a dataset.

  • Generalizing the results of a neural network.

  • Initializing and regularizing a neural network: L1 / L2 regularization, BatchNormalization.

  • Optimization and convergence algorithms.

Standard ML / DL Tools

A simple presentation outlining advantages, disadvantages, position within the ecosystem, and use cases will be provided.

  • Data management tools: Apache Spark, Apache Hadoop Tools.

  • Machine Learning: Numpy, Scipy, Sci-kit.

  • DL high-level frameworks: PyTorch, Keras, Lasagne.

  • Low-level DL frameworks: Theano, Torch, Caffe, Tensorflow.

Convolutional Neural Networks (CNN).

  • Presentation of CNNs: fundamental principles and applications.

  • Basic operation of a CNN: convolutional layer, use of a kernel.

  • Padding & stride, feature map generation, pooling layers. Extensions 1D, 2D, and 3D.

  • Presentation of different CNN architectures that have advanced the state of the art in classification.

  • Images: LeNet, VGG Networks, Network in Network, Inception, Resnet. Presentation of innovations brought about by each architecture and their broader applications (Convolution1x1 or residual connections).

  • Use of an attention model.

  • Application to a common classification case (text or image).

  • CNNs for generation: super-resolution, pixel-to-pixel segmentation. Presentation of

  • Main strategies for increasing feature maps for image generation.

Recurrent Neural Networks (RNN).

  • Presentation of RNNs: fundamental principles and applications.

  • Basic operation of the RNN: hidden activation, back propagation through time, unfolded version.

  • Evolutions towards Gated Recurrent Units (GRUs) and LSTM (Long Short Term Memory).

  • Presentation of the different states and the evolutions brought about by these architectures.

  • Convergence and vanishing gradient problems.

  • Classical architectures: Prediction of a temporal series, classification…

  • RNN Encoder Decoder type architecture. Use of an attention model.

  • NLP applications: word / character encoding, translation.

  • Video Applications: prediction of the next generated image of a video sequence.


Generational models: Variation AutoEncoder (VAE) and Generative Adversarial Networks (GAN).

  • Presentation of generational models, link with CNNs.

  • Auto-encoder: dimensionality reduction and limited generation.

  • Variation Auto-encoder: generational model and approximation of a given distribution. Definition and use of latent space. Reparameterization trick. Applications and limits observed.

  • Generative Adversarial Networks: Fundamentals.

  • Dual Network Architecture (Generator and discriminator) with alternate learning, available cost functions.

  • Convergence of a GAN and difficulties encountered.

  • Improved convergence: Wasserstein GAN, Began. Earth Moving Distance.

  • Applications for image or photograph generation, text generation, super-resolution.

Deep Reinforcement Learning.

  • Presentation of reinforcement learning: controlling an agent in a defined environment.

  • By a state and possible actions.

  • Using a neural network to approximate the state function.

  • Deep Q Learning: experience replay, and application to controlling a video game.

  • Optimizing the learning policy. On-policy && off-policy. Actor-critic architecture. A3C.

  • Applications: controlling a single video game or a digital system.

Part 2 – Theano for Deep Learning

Theano Basics

  • Introduction.

  • Installation and Configuration.

Theano Functions

  • inputs, outputs, updates, givens.

Training and Optimization of a neural network using Theano

  • Neural Network Modeling.

  • Logistic Regression.

  • Hidden Layers.

  • Training a network.

  • Computing and Classification.

  • Optimization.

  • Log Loss.

Testing the model


Part 3 – DNN using Tensorflow

TensorFlow Basics

  • Creation, Initializing, Saving, and Restoring TensorFlow variables.

  • Feeding, Reading and Preloading TensorFlow Data.

  • How to use TensorFlow infrastructure to train models at scale.

  • Visualizing and Evaluating models with TensorBoard.

TensorFlow Mechanics

  • Prepare the Data.

  • Download.

  • Inputs and Placeholders.

  • Build the Graphs.

    • Inference.

    • Loss.

    • Training.

  • Train the Model.

    • The Graph.

    • The Session.

    • Train Loop.

  • Evaluate the Model.

    • Build the Eval Graph.

    • Eval Output.

The Perceptron

  • Activation functions.

  • The perceptron learning algorithm.

  • Binary classification with the perceptron.

  • Document classification with the perceptron.

  • Limitations of the perceptron.

From the Perceptron to Support Vector Machines

  • Kernels and the kernel trick.

  • Maximum margin classification and support vectors.

Artificial Neural Networks

  • Nonlinear decision boundaries.

  • Feedforward and feedback artificial neural networks.

  • Multilayer perceptrons.

  • Minimizing the cost function.

  • Forward propagation.

  • Back propagation.

  • Improving the way neural networks learn.

Convolutional Neural Networks

  • Goals.

  • Model Architecture.

  • Principles.

  • Code Organization.

  • Launching and Training the Model.

  • Evaluating a Model.

Basic Introductions to be given to the below modules (Brief Introduction to be provided based on time availability):

Tensorflow - Advanced Usage

  • Threading and Queues.

  • Distributed TensorFlow.

  • Writing Documentation and Sharing your Model.

  • Customizing Data Readers.

  • Manipulating TensorFlow Model Files.


TensorFlow Serving

  • Introduction.

  • Basic Serving Tutorial.

  • Advanced Serving Tutorial.

  • Serving Inception Model Tutorial.

Kurs İçin Gerekli Önbilgiler

Fizik, matematik ve programlama alanlarında geçmişe sahip olmak. Görüntü işleme faaliyetlerine katılım.

Katılımcıların makine öğrenimi kavramları hakkında önceden bilgi sahibi olması ve Python programlama ve kütüphaneleri üzerinde çalışmış olması gerekmektedir.

 35 Saat

Katılımcı Sayısı


Kişi Başına Fiyat

Danışanlarımızın Yorumları (5)

Yaklaşan Etkinlikler

İlgili Kategoriler