Hands-On Mathematics for Deep Learning: build a solid mathematical foundation for training efficient deep neural networks.
Material type: TextPublication details: Birmingham: Packt Publishing, 2020.Description: vii,349pISBN:- 9781838647292
- 006.31015 DAW
Item type | Current library | Collection | Call number | Status | Date due | Barcode | Item holds | |
---|---|---|---|---|---|---|---|---|
Reference Books | Main Library Reference | Reference | 006.31015 DAW (Browse shelf(Opens below)) | Available | 016073 |
Browsing Main Library shelves, Shelving location: Reference, Collection: Reference Close shelf browser (Hides shelf browser)
Includes Index
Table of ContentsLinear AlgebraVector CalculusProbability and StatisticsOptimizationGraph TheoryLinear Neural NetworksFeedforward Neural NetworksRegularizationConvolutional Neural NetworksRecurrent Neural NetworksAttention MechanismsGenerative ModelsTransfer and Meta LearningGeometric Deep Learning
The main aim of this book is to make the advanced mathematical background accessible to someone with a programming background. This book will equip the readers with not only deep learning architectures but the mathematics behind them. With this book, you will understand the relevant mathematics that goes behind building deep learning models
A comprehensive guide to getting well-versed with the mathematical techniques for building modern deep learning architectures Key Features Understand linear algebra, calculus, gradient algorithms, and other concepts essential for training deep neural networks Learn the mathematical concepts needed to understand how deep learning models function Use deep learning for solving problems related to vision, image, text, and sequence applications Book Description Most programmers and data scientists struggle with mathematics, having either overlooked or forgotten core mathematical concepts. This book uses Python libraries to help you understand the math required to build deep learning (DL) models. You'll begin by learning about core mathematical and modern computational techniques used to design and implement DL algorithms. This book will cover essential topics, such as linear algebra, eigenvalues and eigenvectors, the singular value decomposition concept, and gradient algorithms, to help you understand how to train deep neural networks. Later chapters focus on important neural networks, such as the linear neural network and multilayer perceptrons, with a primary focus on helping you learn how each model works. As you advance, you will delve into the math used for regularization, multi-layered DL, forward propagation, optimization, and backpropagation techniques to understand what it takes to build full-fledged DL models. Finally, you'll explore CNN, recurrent neural network (RNN), and GAN models and their application. By the end of this book, you'll have built a strong foundation in neural networks and DL mathematical concepts, which will help you to confidently research and build custom models in DL. What you will learn Understand the key mathematical concepts for building neural network models Discover core multivariable calculus concepts Improve the performance of deep learning models using optimization techniques Cover optimization algorithms, from basic stochastic gradient descent (SGD) to the advanced Adam optimizer Understand computational graphs and their importance in DL Explore the backpropagation algorithm to reduce output error Cover DL algorithms such as convolutional neural networks (CNNs), sequence models, and generative adversarial networks (GANs) Who this book is for This book is for data scientists, machine learning developers, aspiring deep learning developers, or anyone who wants to understand the foundation of deep learning by l
There are no comments on this title.