Performance Tuning Deep Learning in Python - A Masterclass
This is a step-by-step course in getting the most out of deep learning models on your own predictive modeling projects.
Development ,Data Science,Python
Lectures -115
Duration -4.5 hours
Lifetime Access
Lifetime Access
30-days Money-Back Guarantee
Get your team access to 10000+ top Tutorials Point courses anytime, anywhere.
Course Description
Deep learning neural networks have become easy to create. However, tuning these models for maximum performance remains something of a challenge for most modelers. This course will teach you how to get results as a machine learning practitioner.
The course starts with an introduction to the problem of overfitting and a tour of regularization techniques. Learn through better configured stochastic gradient descent batch size, loss functions, learning rates, and to avoid exploding gradients via gradient clipping. After that, you’ll learn regularization techniques and reduce overfitting by updating the loss function using techniques such as weight regularization, weight constraints, and activation regularization. Post that, you’ll effectively apply dropout, the addition of noise, and early stopping, and combine the predictions from multiple models.
You’ll also look at ensemble learning techniques and diagnose poor model training and problems such as premature convergence and accelerate the model training process. Then, you’ll combine the predictions from multiple models saved during a single training run with techniques such as horizontal ensembles and snapshot ensembles.
Finally, you’ll diagnose high variance in a final model and improve the average predictive skill.
By the end of this course, you’ll learn different techniques for getting better results with deep learning models.
All the resource files are uploaded on the GitHub repository at https://github.com/PacktPublishing/Performance-Tuning-Deep-Learning-Models-Master-Class
Audience
This course is for developers, machine learning engineers, and data scientists that want to enhance the performance of their deep learning models. This is an intermediate level to advanced level course. It's highly recommended that the learner be proficient in Python, Keras, and machine learning.
A solid foundation in machine learning, deep learning, and Python is required to get better results out of this course. You are also recommended to have the core machine learning libraries in Python.
Goals
- Introduction to the problem of overfitting and regularization techniques
- Look at stochastic gradient descent batch size, and other concepts
- Learn to combat overfitting and an introduction of regularization techniques
- Reduce overfitting by updating the loss function using techniques
- Effectively apply dropout, the addition of noise, and early stopping
- Diagnose high variance in a final model and improve average predictive skill
Curriculum
Check out the detailed breakdown of what’s inside the course
Introduction to the Course
12 Lectures
- Introduction 01:54 01:54
- Course Overview 01:49 01:49
- Is This Course Right for You? 01:07 01:07
- Course Structure 01:08 01:08
- Neural Network Defined 02:56 02:56
- Framework for Optional Learning 02:15 02:15
- Optimal Generalization Techniques 02:53 02:53
- Optimal Prediction Techniques 03:27 03:27
- Framework Application 02:56 02:56
- Diagnostic Learning Curves 02:56 02:56
- The Fit of the Model 02:56 02:56
- Unrepresentative Dataset 01:49 01:49
Optimal Learning
54 Lectures
Optimal Generalization
28 Lectures
Optimal Predictions
21 Lectures
Instructor Details
Packt Publishing
Course Certificate
Use your certificate to make a career change or to advance in your current career.
Our students work
with the Best
Related Video Courses
View MoreAnnual Membership
Become a valued member of Tutorials Point and enjoy unlimited access to our vast library of top-rated Video Courses
Subscribe nowOnline Certifications
Master prominent technologies at full length and become a valued certified professional.
Explore Now