首页
学习
活动
专区
工具
TVP
发布
精选内容/技术社群/优惠产品,尽在小程序
立即前往

《动手学深度学习》例子的PyTorch实现

d2l-pytorch

Github项目链接:

This project reproduces the bookDive Into Deep Learning, adapting the code from MXNet into PyTorch.

This project is adapted from the original Dive Into Deep Learning book by Aston Zhang, Zachary C. Lipton, Mu Li, Alex J. Smola and all the community contributors. We have made an effort to modify the book and convert the MXnet code snippets into PyTorch.

Note: Some ipynb notebooks may not be rendered perfectly in Github. We suggest the repo or using nbviewer to view the notebooks.

Contributing

Please feel free to open a Pull Request to contribute a notebook in PyTorch for the rest of the chapters. Before starting out with the notebook, open an issue with the name of the notebook in order to contribute for the same. We will assign that issue to you (if no one has been assigned earlier).

Strictly follow the naming conventions for the IPython Notebooks and the subsections.

Also, if you think there's any section that requires more/better explanation, please use the issue tracker to open an issue and let us know about the same. We'll get back as soon as possible.

Find some code that needs improvement and submit a pull request.

Find a reference that we missed and submit a pull request.

Try not to submit huge pull requests since this makes them hard to understand and incorporate. Better send several smaller ones.

Support

If you like this repo and find it useful, please consider (★) starring it, so that it can reach a broader audience.

References

[1] Original Book Dive Into Deep Learning -> Github Repo

[2] Deep Learning - The Straight Dope

[3] PyTorch - MXNet Cheatsheet

Cite

If you use this work or code for your research please cite the original book with the following bibtex entry.

Chapters

Ch02 Installation

Installation

Ch03 Introduction

Introduction

Ch04 The Preliminaries: A Crashcourse

4.1 Data Manipulation

4.2 Linear Algebra

4.3 Automatic Differentiation

4.4 Probability and Statistics

4.5 Naive Bayes Classification

4.6 Documentation

Ch05 Linear Neural Networks

5.1 Linear Regression

5.2 Linear Regression Implementation from Scratch

5.3 Concise Implementation of Linear Regression

5.4 Softmax Regression

5.5 Image Classification Data (Fashion-MNIST)

5.6 Implementation of Softmax Regression from Scratch

5.7 Concise Implementation of Softmax Regression

Ch06 Multilayer Perceptrons

6.1 Multilayer Perceptron

6.2 Implementation of Multilayer Perceptron from Scratch

6.3 Concise Implementation of Multilayer Perceptron

6.4 Model Selection Underfitting and Overfitting

6.5 Weight Decay

6.6 Dropout

6.7 Forward Propagation Backward Propagation and Computational Graphs

6.8 Numerical Stability and Initialization

6.9 Considering the Environment

6.10 Predicting House Prices on Kaggle

Ch07 Deep Learning Computation

7.1 Layers and Blocks

7.2 Parameter Management

7.3 Deferred Initialization

7.4 Custom Layers

7.5 File I/O

7.6 GPUs

Ch08 Convolutional Neural Networks

8.1 From Dense Layers to Convolutions

8.2 Convolutions for Images

8.3 Padding and Stride

8.4 Multiple Input and Output Channels

8.5 Pooling

8.6 Convolutional Neural Networks (LeNet)

Ch09 Modern Convolutional Networks

9.1 Deep Convolutional Neural Networks (AlexNet)

9.2 Networks Using Blocks (VGG)

9.3 Network in Network (NiN)

9.4 Networks with Parallel Concatenations (GoogLeNet)

9.5 Batch Normalization

9.6 Residual Networks (ResNet)

9.7 Densely Connected Networks (DenseNet)

Ch10 Recurrent Neural Networks

10.1 Sequence Models

10.2 Language Models

10.3 Recurrent Neural Networks

10.4 Text Preprocessing

10.5 Implementation of Recurrent Neural Networks from Scratch

10.6 Concise Implementation of Recurrent Neural Networks

10.7 Backpropagation Through Time

10.8 Gated Recurrent Units (GRU)

10.9 Long Short Term Memory (LSTM)

10.10 Deep Recurrent Neural Networks

10.11 Bidirectional Recurrent Neural Networks

10.12 Machine Translation and DataSets

10.13 Encoder-Decoder Architecture

10.14 Sequence to Sequence

10.15 Beam Search

Ch11 Attention Mechanism

11.1 Attention Mechanism

11.2 Sequence to Sequence with Attention Mechanism

11.3 Transformer

Ch12 Optimization Algorithms

12.1 Optimization and Deep Learning

12.2 Convexity

12.3 Gradient Descent

12.4 Stochastic Gradient Descent

12.5 Mini-batch Stochastic Gradient Descent

12.6 Momentum

12.7 Adagrad

12.8 RMSProp

12.9 Adadelta

12.10 Adam

  • 发表于:
  • 原文链接https://kuaibao.qq.com/s/20190903A09VAN00?refer=cp_1026
  • 腾讯「腾讯云开发者社区」是腾讯内容开放平台帐号(企鹅号)传播渠道之一,根据《腾讯内容开放平台服务协议》转载发布内容。
  • 如有侵权,请联系 cloudcommunity@tencent.com 删除。

扫码

添加站长 进交流群

领取专属 10元无门槛券

私享最新 技术干货

扫码加入开发者社群
领券