提交 c1c609be 编写于 作者: bruce66666's avatar bruce66666

deep learning nlp pytorch

上级 730c9af4
......@@ -4,11 +4,12 @@
**Author**: [Robert Guthrie](https://github.com/rguthrie3/DeepLearningForNLPInPytorch)
This tutorial will walk you through the key ideas of deep learning programming using Pytorch. Many of the concepts (such as the computation graph abstraction and autograd) are not unique to Pytorch and are relevant to any deep learning toolkit out there.
本文带您进入使用pytorch框架进行深度学习编程的核心思想。Pytorch的很多概念(比如计算图抽象和自动求导)并非它所独有的,和其他深度学习框架相关。
我写这篇教程是专门针对那些从未用任何深度学习框架(例如:Tensorflow, Theano, Keras, Dynet)编写代码而从事NLP领域的人。我假设你已经知道NLP领域要解决的核心问题:词性标注、语言模型等等。我也认为你通过[AI](http://aima.cs.berkeley.edu/)这本书中所讲的知识熟悉了神经网络达到了入门的级别.
通常这些课程都会介绍反向传播算法和前馈神经网络,并指出它们是线性组合和非线性组合构成的链。本文在假设你已经有了这些知识的情况下,教你如何开始写深度学习代码。
I am writing this tutorial to focus specifically on NLP for people who have never written code in any deep learning framework (e.g, TensorFlow, Theano, Keras, Dynet). It assumes working knowledge of core NLP problems: part-of-speech tagging, language modeling, etc. It also assumes familiarity with neural networks at the level of an intro AI class (such as one from the Russel and Norvig book). Usually, these courses cover the basic backpropagation algorithm on feed-forward neural networks, and make the point that they are chains of compositions of linearities and non-linearities. This tutorial aims to get you started writing deep learning code, given you have this prerequisite knowledge.
注意这篇文章主要关于_models_,而不是数据。对于所有的模型,我只创建一些数据维度较小的测试示例以便你可以看到权重在训练过程中如何变化。如果你想要尝试一些真实数据,您有能力删除本示例中的模型并重新训练他们。
Note this is about _models_, not data. For all of the models, I just create a few test examples with small dimensionality so you can see how the weights change as it trains. If you have some real data you want to try, you should be able to rip out any of the models from this notebook and use them on it.
![https://pytorch.org/tutorials/_images/sphx_glr_pytorch_tutorial_thumb.png](img/ad16fc851a032d82abda756c6d96f5a6.jpg)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册