舉報

會員
Deep Learning Essentials
Aspiringdatascientistsandmachinelearningexpertswhohavelimitedornoexposuretodeeplearningwillfindthisbooktobeveryuseful.Ifyouarelookingforaresourcethatgetsyouupandrunningwiththefundamentalsofdeeplearningandneuralnetworks,thisbookisforyou.AsthemodelsinthebookaretrainedusingthepopularPython-basedlibrariessuchasTensorflowandKeras,itwouldbeusefultohavesoundprogrammingknowledgeofPython.
目錄(274章)
倒序
- coverpage
- Title Page
- Packt Upsell
- Why subscribe?
- PacktPub.com
- Contributors
- About the authors
- About the reviewer
- Packt is searching for authors like you
- Preface
- Who this book is for
- What this book covers
- To get the most out of this book
- Download the example code files
- Download the color images
- Conventions used
- Get in touch
- Reviews
- Why Deep Learning?
- What is AI and deep learning?
- The history and rise of deep learning
- Why deep learning?
- Advantages over traditional shallow methods
- Impact of deep learning
- The motivation of deep architecture
- The neural viewpoint
- The representation viewpoint
- Distributed feature representation
- Hierarchical feature representation
- Applications
- Lucrative applications
- Success stories
- Deep learning for business
- Future potential and challenges
- Summary
- Getting Yourself Ready for Deep Learning
- Basics of linear algebra
- Data representation
- Data operations
- Matrix properties
- Deep learning with GPU
- Deep learning hardware guide
- CPU cores
- CPU cache size
- RAM size
- Hard drive
- Cooling systems
- Deep learning software frameworks
- TensorFlow – a deep learning library
- Caffe
- MXNet
- Torch
- Theano
- Microsoft Cognitive Toolkit
- Keras
- Framework comparison
- Setting up deep learning on AWS
- Setup from scratch
- Setup using Docker
- Summary
- Getting Started with Neural Networks
- Multilayer perceptrons
- The input layer
- The output layer
- Hidden layers
- Activation functions
- Sigmoid or logistic function
- Tanh or hyperbolic tangent function
- ReLU
- Leaky ReLU and maxout
- Softmax
- Choosing the right activation function
- How a network learns
- Weight initialization
- Forward propagation
- Backpropagation
- Calculating errors
- Backpropagation
- Updating the network
- Automatic differentiation
- Vanishing and exploding gradients
- Optimization algorithms
- Regularization
- Deep learning models
- Convolutional Neural Networks
- Convolution
- Pooling/subsampling
- Fully connected layer
- Overall
- Restricted Boltzmann Machines
- Energy function
- Encoding and decoding
- Contrastive divergence (CD-k)
- Stacked/continuous RBM
- RBM versus Boltzmann Machines
- Recurrent neural networks (RNN/LSTM)
- Cells in RNN and unrolling
- Backpropagation through time
- Vanishing gradient and LTSM
- Cells and gates in LTSM
- Step 1 – The forget gate
- Step 2 – Updating memory/cell state
- Step 3 – The output gate
- Practical examples
- TensorFlow setup and key concepts
- Handwritten digits recognition
- Summary
- Deep Learning in Computer Vision
- Origins of CNNs
- Convolutional Neural Networks
- Data transformations
- Input preprocessing
- Data augmentation
- Network layers
- Convolution layer
- Pooling or subsampling layer
- Fully connected or dense layer
- Network initialization
- Regularization
- Loss functions
- Model visualization
- Handwritten digit classification example
- Fine-tuning CNNs
- Popular CNN architectures
- AlexNet
- Visual Geometry Group
- GoogLeNet
- ResNet
- Summary
- NLP - Vector Representation
- Traditional NLP
- Bag of words
- Weighting the terms tf-idf
- Deep learning NLP
- Motivation and distributed representation
- Word embeddings
- Idea of word embeddings
- Advantages of distributed representation
- Problems of distributed representation
- Commonly used pre-trained word embeddings
- Word2Vec
- Basic idea of Word2Vec
- The word windows
- Generating training data
- Negative sampling
- Hierarchical softmax
- Other hyperparameters
- Skip-Gram model
- The input layer
- The hidden layer
- The output layer
- The loss function
- Continuous Bag-of-Words model
- Training a Word2Vec using TensorFlow
- Using existing pre-trained Word2Vec embeddings
- Word2Vec from Google News
- Using the pre-trained Word2Vec embeddings
- Understanding GloVe
- FastText
- Applications
- Example use cases
- Fine-tuning
- Summary
- Advanced Natural Language Processing
- Deep learning for text
- Limitations of neural networks
- Recurrent neural networks
- RNN architectures
- Basic RNN model
- Training RNN is tough
- Long short-term memory network
- LSTM implementation with tensorflow
- Applications
- Language modeling
- Sequence tagging
- Machine translation
- Seq2Seq inference
- Chatbots
- Summary
- Multimodality
- What is multimodality learning?
- Challenges of multimodality learning
- Representation
- Translation
- Alignment
- Fusion
- Co-learning
- Image captioning
- Show and tell
- Encoder
- Decoder
- Training
- Testing/inference
- Beam Search
- Other types of approaches
- Datasets
- Evaluation
- BLEU
- ROUGE
- METEOR
- CIDEr
- SPICE
- Rank position
- Attention models
- Attention in NLP
- Attention in computer vision
- The difference between hard attention and soft attention
- Visual question answering
- Multi-source based self-driving
- Summary
- Deep Reinforcement Learning
- What is reinforcement learning (RL)?
- Problem setup
- Value learning-based algorithms
- Policy search-based algorithms
- Actor-critic-based algorithms
- Deep reinforcement learning
- Deep Q-network (DQN)
- Experience replay
- Target network
- Reward clipping
- Double-DQN
- Prioritized experience delay
- Dueling DQN
- Implementing reinforcement learning
- Simple reinforcement learning example
- Reinforcement learning with Q-learning example
- Summary
- Deep Learning Hacks
- Massaging your data
- Data cleaning
- Data augmentation
- Data normalization
- Tricks in training
- Weight initialization
- All-zero
- Random initialization
- ReLU initialization
- Xavier initialization
- Optimization
- Learning rate
- Mini-batch
- Clip gradients
- Choosing the loss function
- Multi-class classification
- Multi-class multi-label classification
- Regression
- Others
- Preventing overfitting
- Batch normalization
- Dropout
- Early stopping
- Fine-tuning
- Fine-tuning
- When to use fine-tuning
- When not to use fine-tuning
- Tricks and techniques
- List of pre-trained models
- Model compression
- Summary
- Deep Learning Trends
- Recent models for deep learning
- Generative Adversarial Networks
- Capsule networks
- Novel applications
- Genomics
- Predictive medicine
- Clinical imaging
- Lip reading
- Visual reasoning
- Code synthesis
- Summary
- Other Books You May Enjoy
- Leave a review – let other readers know what you think 更新時間:2021-06-30 19:18:45
推薦閱讀
- PPT,要你好看
- Mastering Matplotlib 2.x
- 空間機器人遙操作系統及控制
- 網上生活必備
- 大數據改變世界
- Arduino &樂高創意機器人制作教程
- Windows游戲程序設計基礎
- TensorFlow Reinforcement Learning Quick Start Guide
- FPGA/CPLD應用技術(Verilog語言版)
- INSTANT Puppet 3 Starter
- 簡明學中文版Photoshop
- 重估:人工智能與賦能社會
- 貫通Hibernate開發
- Moodle 2.0 Course Conversion(Second Edition)
- 智能小車機器人制作大全(第2版)
- 多傳感器數據智能融合理論與應用
- 時序大數據平臺TDengine核心原理與實戰
- 信息技術基礎應用
- 網絡互聯組網配置技術
- 嵌入式系統原理與應用設計
- 單片機應用技能操作和學習指導
- Mastering Microsoft Dynamics CRM 2016
- 零基礎學三菱PLC編程:入門、提高、應用、實例
- Visual Basic.NET+SQL Server全程指南
- 數據庫原理、應用與開發
- 機器學習算法實踐:推薦系統的協同過濾理論及其應用
- Excel 2007表格應用百練成精
- Hands-On Exploratory Data Analysis with R
- Red Hat Enterprise Linux 6從入門到精通
- Learning Alteryx