MARC details
000 -LEADER |
fixed length control field |
11702nam a22002057a 4500 |
005 - DATE AND TIME OF LATEST TRANSACTION |
control field |
20250105141236.0 |
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION |
fixed length control field |
250105b |||||||| |||| 00| 0 eng d |
020 ## - INTERNATIONAL STANDARD BOOK NUMBER |
International Standard Book Number |
9780137470358 |
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER |
Classification number |
006.3 |
Item number |
EKM |
100 ## - MAIN ENTRY--PERSONAL NAME |
Personal name |
Ekman, Magnus |
245 ## - TITLE STATEMENT |
Title |
Learning deep learning: |
Remainder of title |
theory and practice of neural networks, computer vision, natural language processing, and transformers using tensorflow |
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT) |
Name of publisher, distributor, etc. |
Pearson |
Place of publication, distribution, etc. |
Boston |
Date of publication, distribution, etc. |
2022 |
300 ## - PHYSICAL DESCRIPTION |
Extent |
liii, 688 p. |
365 ## - TRADE PRICE |
Price type code |
USD |
Price amount |
74.99 |
500 ## - GENERAL NOTE |
General note |
Table of content:<br/>Foreword by Dr. Anima Anandkumar xxi<br/>Foreword by Dr. Craig Clawson xxiii<br/>Preface xxv<br/>Acknowledgments li<br/>About the Author liii<br/><br/><br/>Chapter 1: The Rosenblatt Perceptron 1<br/><br/>Example of a Two-Input Perceptron 4<br/>The Perceptron Learning Algorithm 7<br/>Limitations of the Perceptron 15<br/>Combining Multiple Perceptrons 17<br/>Implementing Perceptrons with Linear Algebra 20<br/>Geometric Interpretation of the Perceptron 30<br/>Understanding the Bias Term 33<br/>Concluding Remarks on the Perceptron 34<br/><br/>Chapter 2: Gradient-Based Learning 37<br/><br/>Intuitive Explanation of the Perceptron Learning Algorithm 37<br/>Derivatives and Optimization Problems 41<br/>Solving a Learning Problem with Gradient Descent 44<br/>Constants and Variables in a Network 48<br/>Analytic Explanation of the Perceptron Learning Algorithm 49<br/>Geometric Description of the Perceptron Learning Algorithm 51<br/>Revisiting Different Types of Perceptron Plots 52<br/>Using a Perceptron to Identify Patterns 54<br/>Concluding Remarks on Gradient-Based Learning 57<br/><br/>Chapter 3: Sigmoid Neurons and Backpropagation 59<br/><br/>Modified Neurons to Enable Gradient Descent for Multilevel Networks 60<br/>Which Activation Function Should We Use? 66<br/>Function Composition and the Chain Rule 67<br/>Using Backpropagation to Compute the Gradient 69<br/>Backpropagation with Multiple Neurons per Layer 81<br/>Programming Example: Learning the XOR Function 82<br/>Network Architectures 87<br/>Concluding Remarks on Backpropagation 89<br/><br/>Chapter 4: Fully Connected Networks Applied to Multiclass Classification 91<br/><br/>Introduction to Datasets Used When Training Networks 92<br/>Training and Inference 100<br/>Extending the Network and Learning Algorithm to Do Multiclass Classification 101<br/>Network for Digit Classification 102<br/>Loss Function for Multiclass Classification 103<br/>Programming Example: Classifying Handwritten Digits 104<br/>Mini-Batch Gradient Descent 114<br/>Concluding Remarks on Multiclass Classification 115<br/><br/>Chapter 5: Toward DL: Frameworks and Network Tweaks 117<br/><br/>Programming Example: Moving to a DL Framework 118<br/>The Problem of Saturated Neurons and Vanishing Gradients 124<br/>Initialization and Normalization Techniques to Avoid Saturated Neurons 126<br/>Cross-Entropy Loss Function to Mitigate Effect of Saturated Output Neurons 130<br/>Different Activation Functions to Avoid Vanishing Gradient in Hidden Layers 136<br/>Variations on Gradient Descent to Improve Learning 141<br/>Experiment: Tweaking Network and Learning Parameters 143<br/>Hyperparameter Tuning and Cross-Validation 146<br/>Concluding Remarks on the Path Toward Deep Learning 150<br/><br/>Chapter 6: Fully Connected Networks Applied to Regression 153<br/><br/>Output Units 154<br/>The Boston Housing Dataset 160<br/>Programming Example: Predicting House Prices with a DNN 161<br/>Improving Generalization with Regularization 166<br/>Experiment: Deeper and Regularized Models for House Price Prediction 169<br/>Concluding Remarks on Output Units and Regression Problems 170<br/><br/>Chapter 7: Convolutional Neural Networks Applied to Image Classification 171<br/><br/>The CIFAR-10 Dataset 173<br/>Characteristics and Building Blocks for Convolutional Layers 175<br/>Combining Feature Maps into a Convolutional Layer 180<br/>Combining Convolutional and Fully Connected Layers into a Network 181<br/>Effects of Sparse Connections and Weight Sharing 185<br/>Programming Example: Image Classification with a Convolutional Network 190<br/>Concluding Remarks on Convolutional Networks 201<br/><br/>Chapter 8: Deeper CNNs and Pretrained Models 205<br/><br/>VGGNet 206<br/>GoogLeNet 210<br/>ResNet 215<br/>Programming Example: Use a Pretrained ResNet Implementation 223<br/>Transfer Learning 226<br/>Backpropagation for CNN and Pooling 228<br/>Data Augmentation as a Regularization Technique 229<br/>Mistakes Made by CNNs 231<br/>Reducing Parameters with Depthwise Separable Convolutions 232<br/>Striking the Right Network Design Balance with EfficientNet 234<br/>Concluding Remarks on Deeper CNNs 235<br/><br/>Chapter 9: Predicting Time Sequences with Recurrent Neural Networks 237<br/><br/>Limitations of Feedforward Networks 241<br/>Recurrent Neural Networks 242<br/>Mathematical Representation of a Recurrent Layer 243<br/>Combining Layers into an RNN 245<br/>Alternative View of RNN and Unrolling in Time 246<br/>Backpropagation Through Time 248<br/>Programming Example: Forecasting Book Sales 250<br/>Dataset Considerations for RNNs 264<br/>Concluding Remarks on RNNs 265<br/><br/>Chapter 10: Long Short-Term Memory 267<br/><br/>Keeping Gradients Healthy 267<br/>Introduction to LSTM 272<br/>LSTM Activation Functions 277<br/>Creating a Network of LSTM Cells 278<br/>Alternative View of LSTM 280<br/>Related Topics: Highway Networks and Skip Connections 282<br/>Concluding Remarks on LSTM 282<br/><br/>Chapter 11: Text Autocompletion with LSTM and Beam Search 285<br/><br/>Encoding Text 285<br/>Longer-Term Prediction and Autoregressive Models 287<br/>Beam Search 289<br/>Programming Example: Using LSTM for Text Autocompletion 291<br/>Bidirectional RNNs 298<br/>Different Combinations of Input and Output Sequences 300<br/>Concluding Remarks on Text Autocompletion with LSTM 302<br/><br/>Chapter 12: Neural Language Models and Word Embeddings 303<br/><br/>Introduction to Language Models and Their Use Cases 304<br/>Examples of Different Language Models 307<br/>Benefit of Word Embeddings and Insight into How They Work 313<br/>Word Embeddings Created by Neural Language Models 315<br/>Programming Example: Neural Language Model and Resulting Embeddings 319<br/>King − Man + Woman! = Queen 329<br/>King − Man + Woman ! = Queen 331<br/>Language Models, Word Embeddings, and Human Biases 332<br/>Related Topic: Sentiment Analysis of Text 334<br/>Concluding Remarks on Language Models and Word Embeddings 342<br/><br/>Chapter 13: Word Embeddings from word2vec and GloVe 343<br/><br/>Using word2vec to Create Word Embeddings Without a Language Model 344<br/>Additional Thoughts on word2vec 352<br/>word2vec in Matrix Form 353<br/>Wrapping Up word2vec 354<br/>Programming Example: Exploring Properties of GloVe Embeddings 356<br/>Concluding Remarks on word2vec and GloVe 361<br/><br/>Chapter 14: Sequence-to-Sequence Networks and Natural Language Translation 363<br/><br/>Encoder-Decoder Model for Sequence-to-Sequence Learning 366<br/>Introduction to the Keras Functional API 368<br/>Programming Example: Neural Machine Translation 371<br/>Experimental Results 387<br/>Properties of the Intermediate Representation 389<br/>Concluding Remarks on Language Translation 391<br/><br/>Chapter 15: Attention and the Transformer 393<br/><br/>Rationale Behind Attention 394<br/>Attention in Sequence-to-Sequence Networks 395<br/>Alternatives to Recurrent Networks 406<br/>Self-Attention 407<br/>Multi-head Attention 410<br/>The Transformer 411<br/>Concluding Remarks on the Transformer 415<br/><br/>Chapter 16: One-to-Many Network for Image Captioning 417<br/><br/>Extending the Image Captioning Network with Attention 420<br/>Programming Example: Attention-Based Image Captioning 421<br/>Concluding Remarks on Image Captioning 443<br/><br/>Chapter 17: Medley of Additional Topics 447<br/><br/>Autoencoders 448<br/>Multimodal Learning 459<br/>Multitask Learning 469<br/>Process for Tuning a Network 477<br/>Neural Architecture Search 482<br/>Concluding Remarks 502<br/><br/>Chapter 18: Summary and Next Steps 503<br/><br/>Things You Should Know by Now 503<br/>Ethical AI and Data Ethics 505<br/>Things You Do Not Yet Know 512<br/>Next Steps 516<br/><br/>Appendix A: Linear Regression and Linear Classifiers 519<br/><br/>Linear Regression as a Machine Learning Algorithm 519<br/>Computing Linear Regression Coefficients 523<br/>Classification with Logistic Regression 525<br/>Classifying XOR with a Linear Classifier 528<br/>Classification with Support Vector Machines 531<br/>Evaluation Metrics for a Binary Classifier 533<br/><br/>Appendix B: Object Detection and Segmentation 539<br/><br/>Object Detection 540<br/>Semantic Segmentation 549<br/>Instance Segmentation with Mask R-CNN 559<br/><br/>Appendix C: Word Embeddings Beyond word2vec and GloVe 563<br/><br/>Wordpieces 564<br/>FastText 566<br/>Character-Based Method 567<br/>ELMo 572<br/>Related Work 575<br/><br/>Appendix D: GPT, BERT, AND RoBERTa 577<br/><br/>GPT 578<br/>BERT 582<br/>RoBERTa 586<br/>Historical Work Leading Up to GPT and BERT 588<br/>Other Models Based on the Transformer 590<br/><br/>Appendix E: Newton-Raphson versus Gradient Descent 593<br/><br/>Newton-Raphson Root-Finding Method 594<br/>Relationship Between Newton-Raphson and Gradient Descent 597<br/><br/>Appendix F: Matrix Implementation of Digit Classification Network 599<br/><br/>Single Matrix 599<br/>Mini-Batch Implementation 602<br/><br/>Appendix G: Relating Convolutional Layers to Mathematical Convolution 607Appendix H: Gated Recurrent Units 613<br/><br/>Alternative GRU Implementation 616<br/>Network Based on the GRU 616<br/><br/>Appendix I: Setting up a Development Environment 621<br/><br/>Python 622<br/>Programming Environment 623<br/>Programming Examples 624<br/>Datasets 625<br/>Installing a DL Framework 628<br/>TensorFlow Specific Considerations 630<br/>Key Differences Between PyTorch and TensorFlow 631<br/><br/>Appendix J: Cheat Sheets 637<br/><br/><br/>Works Cited 647<br/>Index 667<br/>[https://www.pearson.com/en-us/subject-catalog/p/learning-deep-learning-theory-and-practice-of-neural-networks-computer-vision-natural-language-processing-and-transformers-using-tensorflow/P200000009457/9780137470358?tab=table-of-contents] |
520 ## - SUMMARY, ETC. |
Summary, etc. |
NVIDIA's Full-Color Guide to Deep Learning: All StudentsNeed to Get Started and Get Results<br/>Learning Deep Learning is a complete guide to DL. Illuminating both the core concepts and the hands-on programming techniques needed to succeed, this text can be used for students with prior programming experince but with no prior machine learning or statistics experience.<br/><br/>After introducing the essential building blocks of deep neural networks, such as artificial neurons and fully connected, convolutional, and recurrent layers, Ekman shows how to use them to build advanced architectures, including the Transformer. He describes how these concepts are used to build modern networks for computer vision and natural language processing (NLP), including Mask R-CNN, GPT, and BERT. And he explains a natural language translator and a system generating natural language descriptions of images.<br/>Throughout, Ekman provides concise, well-annotated code examples using TensorFlow with Keras. Corresponding PyTorch examples are provided online, and the book thereby covers the two dominating Python libraries for DL used in industry and academia. He concludes with an introduction to neural architecture search (NAS), exploring important ethical issues and providing resources for further learning.<br/><br/>Explore and master core concepts: perceptrons, gradient-based learning, sigmoid neurons, and back propagation<br/>See how DL frameworks make it easier to develop more complicated and useful neural networks<br/>Discover how convolutional neural networks (CNNs) revolutionize image classification and analysis<br/>Apply recurrent neural networks (RNNs) and long short-term memory (LSTM) to text and other variable-length sequences<br/>Master NLP with sequence-to-sequence networks and the Transformer architecture<br/>Build applications for natural language translation and image captioning<br/>(https://www.pearson.com/en-us/subject-catalog/p/learning-deep-learning-theory-and-practice-of-neural-networks-computer-vision-natural-language-processing-and-transformers-using-tensorflow/P200000009457/9780137470358?tab=title-overview) |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name as entry element |
Computer vision |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM |
Topical term or geographic name as entry element |
Machine learning |
942 ## - ADDED ENTRY ELEMENTS (KOHA) |
Koha item type |
Book |
Source of classification or shelving scheme |
Dewey Decimal Classification |