舉報

會員
Deep Learning with PyTorch
Thisbookisformachinelearningengineers,dataanalysts,datascientistsinterestedindeeplearningandarelookingtoexploreimplementingadvancedalgorithmsinPyTorch.Someknowledgeofmachinelearningishelpfulbutnotamandatoryneed.WorkingknowledgeofPythonprogrammingisexpected.
目錄(248章)
倒序
- coverpage
- Title Page
- Dedication
- Packt Upsell
- Why subscribe?
- PacktPub.com
- Foreword
- Contributors
- About the author
- About the reviewer
- Packt is searching for authors like you
- Preface
- Who this book is for
- What this book covers
- To get the most out of this book
- Download the example code files
- Download the color images
- Conventions used
- Get in touch
- Reviews
- Getting Started with Deep Learning Using PyTorch
- Artificial intelligence
- The history of AI
- Machine learning
- Examples of machine learning in real life
- Deep learning
- Applications of deep learning
- Hype associated with deep learning
- The history of deep learning
- Why now?
- Hardware availability
- Data and algorithms
- Deep learning frameworks
- PyTorch
- Summary
- Building Blocks of Neural Networks
- Installing PyTorch
- Our first neural network
- Data preparation
- Scalar (0-D tensors)
- Vectors (1-D tensors)
- Matrix (2-D tensors)
- 3-D tensors
- Slicing tensors
- 4-D tensors
- 5-D tensors
- Tensors on GPU
- Variables
- Creating data for our neural network
- Creating learnable parameters
- Neural network model
- Network implementation
- Loss function
- Optimize the neural network
- Loading data
- Dataset class
- DataLoader class
- Summary
- Diving Deep into Neural Networks
- Deep dive into the building blocks of neural networks
- Layers – fundamental blocks of neural networks
- Non-linear activations
- Sigmoid
- Tanh
- ReLU
- Leaky ReLU
- PyTorch non-linear activations
- The PyTorch way of building deep learning algorithms
- Model architecture for different machine learning problems
- Loss functions
- Optimizing network architecture
- Image classification using deep learning
- Loading data into PyTorch tensors
- Loading PyTorch tensors as batches
- Building the network architecture
- Training the model
- Summary
- Fundamentals of Machine Learning
- Three kinds of machine learning problems
- Supervised learning
- Unsupervised learning
- Reinforcement learning
- Machine learning glossary
- Evaluating machine learning models
- Training validation and test split
- Simple holdout validation
- K-fold validation
- K-fold validation with shuffling
- Data representativeness
- Time sensitivity
- Data redundancy
- Data preprocessing and feature engineering
- Vectorization
- Value normalization
- Handling missing values
- Feature engineering
- Overfitting and underfitting
- Getting more data
- Reducing the size of the network
- Applying weight regularization
- Dropout
- Underfitting
- Workflow of a machine learning project
- Problem definition and dataset creation
- Measure of success
- Evaluation protocol
- Prepare your data
- Baseline model
- Applying regularization
- Learning rate picking strategies
- Summary
- Deep Learning for Computer Vision
- Introduction to neural networks
- MNIST – getting data
- Building a CNN model from scratch
- Conv2d
- Pooling
- Nonlinear activation – ReLU
- View
- Linear layer
- Training the model
- Classifying dogs and cats – CNN from scratch
- Classifying dogs and cats using transfer learning
- Creating and exploring a VGG16 model
- Freezing the layers
- Fine-tuning VGG16
- Training the VGG16 model
- Calculating pre-convoluted features
- Understanding what a CNN model learns
- Visualizing outputs from intermediate layers
- Visualizing weights of the CNN layer
- Summary
- Deep Learning with Sequence Data and Text
- Working with text data
- Tokenization
- Converting text into characters
- Converting text into words
- N-gram representation
- Vectorization
- One-hot encoding
- Word embedding
- Training word embedding by building a sentiment classifier
- Downloading IMDB data and performing text tokenization
- torchtext.data
- torchtext.datasets
- Building vocabulary
- Generate batches of vectors
- Creating a network model with embedding
- Training the model
- Using pretrained word embeddings
- Downloading the embeddings
- Loading the embeddings in the model
- Freeze the embedding layer weights
- Recursive neural networks
- Understanding how RNN works with an example
- LSTM
- Long-term dependency
- LSTM networks
- Preparing the data
- Creating batches
- Creating the network
- Training the model
- Convolutional network on sequence data
- Understanding one-dimensional convolution for sequence data
- Creating the network
- Training the model
- Summary
- Generative Networks
- Neural style transfer
- Loading the data
- Creating the VGG model
- Content loss
- Style loss
- Extracting the losses
- Creating loss function for each layers
- Creating the optimizer
- Training
- Generative adversarial networks
- Deep convolutional GAN
- Defining the generator network
- Transposed convolutions
- Batch normalization
- Generator
- Defining the discriminator network
- Defining loss and optimizer
- Training the discriminator
- Training the discriminator with real images
- Training the discriminator with fake images
- Training the generator network
- Training the complete network
- Inspecting the generated images
- Language modeling
- Preparing the data
- Generating the batches
- Batches
- Backpropagation through time
- Defining a model based on LSTM
- Defining the train and evaluate functions
- Training the model
- Summary
- Modern Network Architectures
- Modern network architectures
- ResNet
- Creating PyTorch datasets
- Creating loaders for training and validation
- Creating a ResNet model
- Extracting convolutional features
- Creating a custom PyTorch dataset class for the pre-convoluted features and loader
- Creating a simple linear model
- Training and validating the model
- Inception
- Creating an Inception model
- Extracting convolutional features using register_forward_hook
- Creating a new dataset for the convoluted features
- Creating a fully connected model
- Training and validating the model
- Densely connected convolutional networks – DenseNet
- DenseBlock
- DenseLayer
- Creating a DenseNet model
- Extracting DenseNet features
- Creating a dataset and loaders
- Creating a fully connected model and train
- Model ensembling
- Creating models
- Extracting the image features
- Creating a custom dataset along with data loaders
- Creating an ensembling model
- Training and validating the model
- Encoder-decoder architecture
- Encoder
- Decoder
- Summary
- What Next?
- What next?
- Overview
- Interesting ideas to explore
- Object detection
- Image segmentation
- OpenNMT in PyTorch
- Alien NLP
- fast.ai – making neural nets uncool again
- Open Neural Network Exchange
- How to keep yourself updated
- Summary
- Other Books You May Enjoy
- Leave a review - let other readers know what you think 更新時間:2021-06-24 19:17:02
推薦閱讀
- Windows phone 7.5 application development with F#
- 深入理解Spring Cloud與實戰
- Instant uTorrent
- Intel FPGA/CPLD設計(高級篇)
- VCD、DVD原理與維修
- 電腦高級維修及故障排除實戰
- Building 3D Models with modo 701
- Machine Learning with Go Quick Start Guide
- 計算機組裝維修與外設配置(高等職業院校教改示范教材·計算機系列)
- Internet of Things Projects with ESP32
- LPC1100系列處理器原理及應用
- 3D Printing Blueprints
- Arduino項目開發:智能生活
- 筆記本電腦維修技能實訓
- Advanced Machine Learning with R
- Arduino案例實戰(卷Ⅳ)
- 超炫的35個Arduino制作項目
- Spring Cloud微服務架構開發
- 新編計算機組裝與維護
- Raspberry Pi Media Center
- Blender 2.6 Cycles:Materials and Textures Cookbook
- 51單片機典型模塊開發查詢手冊
- Machine Learning for Algorithmic Trading
- 機房工程設計施工與管理維護技術
- Axure RP Prototyping Cookbook
- Android Programming with Kotlin for Beginners
- 通用源碼閱讀指導書:MyBatis源碼詳解
- User Experience Mapping
- 多媒體技術應用實訓教程
- STC單片機原理及應用:從器件、匯編、C到操作系統的分析和設計(立體化教程)(第2版)