Learning Neural Networks with Tensorflow

Watch Learning Neural Networks with Tensorflow

  • 2017
  • 1 Season

Learning Neural Networks with Tensorflow from Packt Publishing is an interactive online course that provides in-depth knowledge and hands-on practice on building, training, and deploying neural networks using TensorFlow. In this course, viewers are guided by Roland Meertens, a highly experienced developer and trainer with a profound understanding of the latest TensorFlow updates and applications.

The course is organized into eleven distinct sections and 61 different modules, each designed to cater to the needs of different skill levels, from beginners to experienced developers. The course covers various essential topics, including how to install TensorFlow, create and train different types of neural networks, use data sets to produce optimal results, and apply these concepts in real-world applications. The lessons are structured in an easy-to-follow format, with many examples, visual aids, quizzes, and code snippets to reinforce learning.

The first section of the course provides an introduction to TensorFlow and its usage, followed by the second section, which focuses on the basics of neural networks. Topics covered in this section include understanding the essential concepts of neural networks, the different types of data used in neural networks, and how to build a neural network from scratch.

The third section of the course is all about training a neural network. Here, viewers learn how to prepare data, use TensorFlow to train a model, optimize the weights of the model, and assess the model's performance. The course is designed to enable learners to gain a deep understanding of the inner workings of neural networks, along with practical exposure to developing and training them.

Section four introduces deep learning, where Roland Meertens explains how to use TensorFlow and deep learning techniques to solve complex problems. This section covers topics like convolutional neural networks, recurrent neural networks, long short-term memory, and how to implement these techniques in TensorFlow.

The fifth section is on the topic of computer vision, where neural networks are used to make predictions based on images and videos. This section covers concepts such as identifying objects within images, recognizing faces, and performing image segmentation. The section consists of ten modules and includes code snippets that learners can try out on their own.

In the sixth section, viewers are introduced to natural language processing. Here, Roland Meertens covers various machine learning techniques that can be used to process natural language data like text, audio, articles, blogs, or reviews. The section also introduces viewers to various deep learning models, such as Recurrent Neural Networks and Convolutional Neural Networks, which are used to process and analyze text data.

Section seven of the course explores how to use TensorFlow for time-series data, such as financial data, energy consumption, or weather data. The section covers techniques like LSTM, RNN, and more, and helps learners understand how to use these techniques to forecast future values of the data.

The eighth section covers transfer learning, a technique used to leverage the features learned on one task to enhance the performance of another task. In this section, learners are introduced to different pre-trained models and learn how to fine-tune these models to address their specific problem.

In section nine, viewers learn about generative models, which are used to produce new data samples that match the statistical characteristics of the original data distribution. The section covers topics such as Variational Autoencoders and Generative Adversarial Networks and how to use TensorFlow to generate realistic images, videos, or natural language text.

Section ten covers reinforcement learning, a popular technique used to enable machines to make decisions by interacting with the environment. Roland Meertens explains how to develop decision-making agents using reinforcement learning algorithms like Q-Learning, SARSA, and Value Iteration.

Finally, section eleven deals with deploying TensorFlow models to production environments. The section covers techniques like Docker, Kubernetes, and TensorFlow Serving and helps learners understand how to deploy their models in a scalable and reliable way.

In conclusion, Learning Neural Networks with TensorFlow from Packt Publishing is an excellent course for developers who are interested in building and deploying machine learning models using TensorFlow. With Roland Meertens as the instructor, this course is an excellent resource to learn both the essential concepts behind neural networks and how to implement them using TensorFlow. With its practical approach and easy-to-follow format, the course can help developers from all skill levels master the essential aspects of neural networks and use them in real-world applications.

Learning Neural Networks with Tensorflow is a series that is currently running and has 1 seasons (24 episodes). The series first aired on November 27, 2017.

Filter by Source

Seasons
Understanding What Your Network Learned -Visualizing Activations
25. Understanding What Your Network Learned -Visualizing Activations
November 27, 2017
Neural networks sometimes learn something you don't expect. Looking at activations can be an important tool to verify your network is learning something that makes sense. We will also evaluate the performance of our network by drawing a ROC curve.
Batch Normalization
24. Batch Normalization
November 27, 2017
Each layer learns to respond to the output of the previous layer during backpropagation. A trick to speed up this process AND get better results is called batch normalization. We will add it to the layers in our network.
Building a Convolutional Neural Network
23. Building a Convolutional Neural Network
November 27, 2017
We loaded our data and preprocessed our images. Now it's time to see how well our well known approach of convolutional layers works on this dataset.
Building an Input Pipeline in TensorFlow
22. Building an Input Pipeline in TensorFlow
November 27, 2017
As there are a lot of images, loading them all into our memory would require a lot from our computer space. Instead we will build a pipeline in TensorFlow that reads the images when we need them.
Large-Scale CelebFaces Attributes (CelebA) Dataset
21. Large-Scale CelebFaces Attributes (CelebA) Dataset
November 27, 2017
To analyze the faces of celebrities, we need to have a lot of data. The CelebA dataset contains more than 200, 000 images of celebrities.
Optimization and Loss Functions
20. Optimization and Loss Functions
November 27, 2017
We currently used the mean squared error loss function and normal gradient descent. The softmax cross entropy function performs better for classification functions. We will also look at the momentum and the adam optimizer, which often perform better.
From Activations to Probabilities - the Softmax Function
19. From Activations to Probabilities - the Softmax Function
November 27, 2017
Knowing the output activations of a neural network is great, but often you want to see a "probability" per output class. To do this we introduce the softmax function.
Convolution and Pooling Layers
17. Convolution and Pooling Layers
November 27, 2017
If you move an image, it's still easy for humans to recognize the image. With our dense layers, the network has to "learn" all positions a character can be at. We will introduce convolutional layers and pooling layers to counter this problem
Dense Layer Approach
16. Dense Layer Approach
November 27, 2017
We will apply what we learned in the previous section on these images and build a deep Neural Network with fully connected layers. We will also write an evaluation function that determines the accuracy of a Neural Network.
Images of Written Digits
15. Images of Written Digits
November 27, 2017
We want to download a dataset with images of written digits and save these digits to our datasets folder. We will visualize them with Matplotlib after reshaping them.
The Importance of Hyperparameters
14. The Importance of Hyperparameters
November 27, 2017
There are several better methods to estimate these hyperparameters, and we will try grid-search over the learning rate parameter to improve our performance.
Understanding the Activation Function
13. Understanding the Activation Function
November 27, 2017
In this video, we will explore other activation functions, and will look at how well the network performs the ReLU function which is used as activation function.
Preprocessing Data
12. Preprocessing Data
November 27, 2017
Learn how preprocessing data can give big performance boosts to Neural Networks.
First Approach - Easy Layer Building
11. First Approach - Easy Layer Building
November 27, 2017
Replicate the Neural Network made in the previous section to see how well this works.
Ground State Energies of 16,242 Molecules
10. Ground State Energies of 16,242 Molecules
November 27, 2017
Download the data from Kaggle and see what is in the dataset.
Overfitting - Why We Split Our Train and Test Data
9. Overfitting - Why We Split Our Train and Test Data
November 27, 2017
Although during training it may look as if our neural network learned to classify everything, it's possible it does not generalize to the whole dataset. To see how well our network performs we have to split our data into training and test set.
Backpropagation
8. Backpropagation
November 27, 2017
Determine the error that the network made, and how we can optimize the network to reduce this error.
The Human Brain and How to Formalize It
7. The Human Brain and How to Formalize It
November 27, 2017
Look at the human brain for inspiration on how computers can learn something and learn how to manually design a Neural Network.
The Iris Dataset
6. The Iris Dataset
November 27, 2017
Get the Iris dataset and inspect it. Find insights in how to recognize flowers.
Understanding TensorFlow
5. Understanding TensorFlow
November 27, 2017
Look at the TensorFlow software, and understand what its definition is. The author will build some graphs, and will explain what they do and how to evaluate them in a session. Some TensorFlow functions are compared to their NumPy equivalent.
Our Code, in a Jupyter Notebook
4. Our Code, in a Jupyter Notebook
November 27, 2017
Learn how to download the source code for this course and build the Docker image. The author will show what commands to enter, and how viewers can open the Jupyter Notebook. Finally, he will show viewers what a Jupyter Notebook is and how it works.
Why We Use Docker and Installation Instructions
3. Why We Use Docker and Installation Instructions
November 27, 2017
In this course, the viewers only need to install one thing, and that is Docker. With this tool, we put a "virtual operating system" on the viewer's computer which has all dependencies that they need for this course.
Solving Public Datasets
2. Solving Public Datasets
November 27, 2017
People often discuss whether being good at deep learning requires a lot of knowledge like a scientist, or requires a lot of practice like an artist. You need a combination of both to build state of the art models.
The Course Overview
1. The Course Overview
November 27, 2017
This video provides an overview of the entire course.
Description
Where to Watch Learning Neural Networks with Tensorflow
Learning Neural Networks with Tensorflow is available for streaming on the Packt Publishing website, both individual episodes and full seasons. You can also watch Learning Neural Networks with Tensorflow on demand at Amazon.
  • Premiere Date
    November 27, 2017