舉報

會員
Deep Learning with PyTorch
Thisbookisformachinelearningengineers,dataanalysts,datascientistsinterestedindeeplearningandarelookingtoexploreimplementingadvancedalgorithmsinPyTorch.Someknowledgeofmachinelearningishelpfulbutnotamandatoryneed.WorkingknowledgeofPythonprogrammingisexpected.
目錄(248章)
倒序
- coverpage
- Title Page
- Dedication
- Packt Upsell
- Why subscribe?
- PacktPub.com
- Foreword
- Contributors
- About the author
- About the reviewer
- Packt is searching for authors like you
- Preface
- Who this book is for
- What this book covers
- To get the most out of this book
- Download the example code files
- Download the color images
- Conventions used
- Get in touch
- Reviews
- Getting Started with Deep Learning Using PyTorch
- Artificial intelligence
- The history of AI
- Machine learning
- Examples of machine learning in real life
- Deep learning
- Applications of deep learning
- Hype associated with deep learning
- The history of deep learning
- Why now?
- Hardware availability
- Data and algorithms
- Deep learning frameworks
- PyTorch
- Summary
- Building Blocks of Neural Networks
- Installing PyTorch
- Our first neural network
- Data preparation
- Scalar (0-D tensors)
- Vectors (1-D tensors)
- Matrix (2-D tensors)
- 3-D tensors
- Slicing tensors
- 4-D tensors
- 5-D tensors
- Tensors on GPU
- Variables
- Creating data for our neural network
- Creating learnable parameters
- Neural network model
- Network implementation
- Loss function
- Optimize the neural network
- Loading data
- Dataset class
- DataLoader class
- Summary
- Diving Deep into Neural Networks
- Deep dive into the building blocks of neural networks
- Layers – fundamental blocks of neural networks
- Non-linear activations
- Sigmoid
- Tanh
- ReLU
- Leaky ReLU
- PyTorch non-linear activations
- The PyTorch way of building deep learning algorithms
- Model architecture for different machine learning problems
- Loss functions
- Optimizing network architecture
- Image classification using deep learning
- Loading data into PyTorch tensors
- Loading PyTorch tensors as batches
- Building the network architecture
- Training the model
- Summary
- Fundamentals of Machine Learning
- Three kinds of machine learning problems
- Supervised learning
- Unsupervised learning
- Reinforcement learning
- Machine learning glossary
- Evaluating machine learning models
- Training validation and test split
- Simple holdout validation
- K-fold validation
- K-fold validation with shuffling
- Data representativeness
- Time sensitivity
- Data redundancy
- Data preprocessing and feature engineering
- Vectorization
- Value normalization
- Handling missing values
- Feature engineering
- Overfitting and underfitting
- Getting more data
- Reducing the size of the network
- Applying weight regularization
- Dropout
- Underfitting
- Workflow of a machine learning project
- Problem definition and dataset creation
- Measure of success
- Evaluation protocol
- Prepare your data
- Baseline model
- Large model enough to overfit
- Applying regularization
- Learning rate picking strategies
- Summary
- Deep Learning for Computer Vision
- Introduction to neural networks
- MNIST – getting data
- Building a CNN model from scratch
- Conv2d
- Pooling
- Nonlinear activation – ReLU
- View
- Linear layer
- Training the model
- Classifying dogs and cats – CNN from scratch
- Classifying dogs and cats using transfer learning
- Creating and exploring a VGG16 model
- Freezing the layers
- Fine-tuning VGG16
- Training the VGG16 model
- Calculating pre-convoluted features
- Understanding what a CNN model learns
- Visualizing outputs from intermediate layers
- Visualizing weights of the CNN layer
- Summary
- Deep Learning with Sequence Data and Text
- Working with text data
- Tokenization
- Converting text into characters
- Converting text into words
- N-gram representation
- Vectorization
- One-hot encoding
- Word embedding
- Training word embedding by building a sentiment classifier
- Downloading IMDB data and performing text tokenization
- torchtext.data
- torchtext.datasets
- Building vocabulary
- Generate batches of vectors
- Creating a network model with embedding
- Training the model
- Using pretrained word embeddings
- Downloading the embeddings
- Loading the embeddings in the model
- Freeze the embedding layer weights
- Recursive neural networks
- Understanding how RNN works with an example
- LSTM
- Long-term dependency
- LSTM networks
- Preparing the data
- Creating batches
- Creating the network
- Training the model
- Convolutional network on sequence data
- Understanding one-dimensional convolution for sequence data
- Creating the network
- Training the model
- Summary
- Generative Networks
- Neural style transfer
- Loading the data
- Creating the VGG model
- Content loss
- Style loss
- Extracting the losses
- Creating loss function for each layers
- Creating the optimizer
- Training
- Generative adversarial networks
- Deep convolutional GAN
- Defining the generator network
- Transposed convolutions
- Batch normalization
- Generator
- Defining the discriminator network
- Defining loss and optimizer
- Training the discriminator
- Training the discriminator with real images
- Training the discriminator with fake images
- Training the generator network
- Training the complete network
- Inspecting the generated images
- Language modeling
- Preparing the data
- Generating the batches
- Batches
- Backpropagation through time
- Defining a model based on LSTM
- Defining the train and evaluate functions
- Training the model
- Summary
- Modern Network Architectures
- Modern network architectures
- ResNet
- Creating PyTorch datasets
- Creating loaders for training and validation
- Creating a ResNet model
- Extracting convolutional features
- Creating a custom PyTorch dataset class for the pre-convoluted features and loader
- Creating a simple linear model
- Training and validating the model
- Inception
- Creating an Inception model
- Extracting convolutional features using register_forward_hook
- Creating a new dataset for the convoluted features
- Creating a fully connected model
- Training and validating the model
- Densely connected convolutional networks – DenseNet
- DenseBlock
- DenseLayer
- Creating a DenseNet model
- Extracting DenseNet features
- Creating a dataset and loaders
- Creating a fully connected model and train
- Model ensembling
- Creating models
- Extracting the image features
- Creating a custom dataset along with data loaders
- Creating an ensembling model
- Training and validating the model
- Encoder-decoder architecture
- Encoder
- Decoder
- Summary
- What Next?
- What next?
- Overview
- Interesting ideas to explore
- Object detection
- Image segmentation
- OpenNMT in PyTorch
- Alien NLP
- fast.ai – making neural nets uncool again
- Open Neural Network Exchange
- How to keep yourself updated
- Summary
- Other Books You May Enjoy
- Leave a review - let other readers know what you think 更新時間:2021-06-24 19:17:02
推薦閱讀
- 24小時學會電腦組裝與維護
- Learning SQL Server Reporting Services 2012
- Linux KVM虛擬化架構實戰指南
- 網絡服務器配置與管理(第3版)
- 嵌入式系統設計教程
- 單片機系統設計與開發教程
- 龍芯自主可信計算及應用
- Blender Quick Start Guide
- 電腦組裝與維護即時通
- 筆記本電腦芯片級維修從入門到精通(圖解版)
- 嵌入式系統原理及應用:基于ARM Cortex-M4體系結構
- 多媒體應用技術(第2版)
- 零基礎輕松學修電腦主板
- Machine Learning Projects for Mobile Applications
- Unreal Development Kit Game Programming with UnrealScript:Beginner's Guide
- CPU設計實戰:LoongArch版
- Unreal Engine 4 AI Programming Essentials
- 計算機組裝與維護項目化教程(第二版)
- Hands-On Game Development with WebAssembly
- INSTANT Website Optimization for Retina Displays How-to
- Hands-On Python Deep Learning for the Web
- Learn Human:Computer Interaction
- Hands-On Explainable AI(XAI) with Python
- 計算機組裝與維護
- Ouya Unity Game Development
- 電腦軟硬件維修從入門到精通
- 硅谷之火:個人計算機的誕生與衰落(第3版)
- 計算機組裝與維修技術(第2版)
- MQTT Essentials:A Lightweight IoT Protocol
- KVM實戰:原理、進階與性能調優