mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
Remove HTML tags from README.md (#9296)
Summary: This change makes README.md compatible with both Github and VSTS markdown engines. Images can be reduced if necessary Pull Request resolved: https://github.com/pytorch/pytorch/pull/9296 Differential Revision: D8874931 Pulled By: soumith fbshipit-source-id: 0c530c1e00b06fc891301644c92c33007060bf27
This commit is contained in:
parent
4ff636a3fd
commit
11fc16dc98
40
README.md
40
README.md
|
|
@ -1,4 +1,4 @@
|
||||||
<p align="center"><img width="40%" src="docs/source/_static/img/pytorch-logo-dark.png" /></p>
|

|
||||||
|
|
||||||
--------------------------------------------------------------------------------
|
--------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
|
@ -34,32 +34,14 @@ See also the [ci.pytorch.org HUD](https://ezyang.github.io/pytorch-ci-hud/build/
|
||||||
|
|
||||||
At a granular level, PyTorch is a library that consists of the following components:
|
At a granular level, PyTorch is a library that consists of the following components:
|
||||||
|
|
||||||
<table>
|
| Component | Description |
|
||||||
<tr>
|
| ---- | --- |
|
||||||
<td><b> torch </b></td>
|
| **torch** | a Tensor library like NumPy, with strong GPU support |
|
||||||
<td> a Tensor library like NumPy, with strong GPU support </td>
|
| **torch.autograd** | a tape-based automatic differentiation library that supports all differentiable Tensor operations in torch |
|
||||||
</tr>
|
| **torch.nn** | a neural networks library deeply integrated with autograd designed for maximum flexibility |
|
||||||
<tr>
|
| **torch.multiprocessing** | Python multiprocessing, but with magical memory sharing of torch Tensors across processes. Useful for data loading and Hogwild training |
|
||||||
<td><b> torch.autograd </b></td>
|
| **torch.utils** | DataLoader, Trainer and other utility functions for convenience |
|
||||||
<td> a tape-based automatic differentiation library that supports all differentiable Tensor operations in torch </td>
|
| **torch.legacy(.nn/.optim)** | legacy code that has been ported over from torch for backward compatibility reasons |
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td><b> torch.nn </b></td>
|
|
||||||
<td> a neural networks library deeply integrated with autograd designed for maximum flexibility </td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td><b> torch.multiprocessing </b></td>
|
|
||||||
<td> Python multiprocessing, but with magical memory sharing of torch Tensors across processes. Useful for data loading and Hogwild training. </td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td><b> torch.utils </b></td>
|
|
||||||
<td> DataLoader, Trainer and other utility functions for convenience </td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td><b> torch.legacy(.nn/.optim) </b></td>
|
|
||||||
<td> legacy code that has been ported over from torch for backward compatibility reasons </td>
|
|
||||||
</tr>
|
|
||||||
</table>
|
|
||||||
|
|
||||||
Usually one uses PyTorch either as:
|
Usually one uses PyTorch either as:
|
||||||
|
|
||||||
|
|
@ -72,7 +54,7 @@ Elaborating further:
|
||||||
|
|
||||||
If you use NumPy, then you have used Tensors (a.k.a ndarray).
|
If you use NumPy, then you have used Tensors (a.k.a ndarray).
|
||||||
|
|
||||||
<p align=center><img width="30%" src="docs/source/_static/img/tensor_illustration.png" /></p>
|

|
||||||
|
|
||||||
PyTorch provides Tensors that can live either on the CPU or the GPU, and accelerate
|
PyTorch provides Tensors that can live either on the CPU or the GPU, and accelerate
|
||||||
compute by a huge amount.
|
compute by a huge amount.
|
||||||
|
|
@ -99,7 +81,7 @@ from several research papers on this topic, as well as current and past work suc
|
||||||
While this technique is not unique to PyTorch, it's one of the fastest implementations of it to date.
|
While this technique is not unique to PyTorch, it's one of the fastest implementations of it to date.
|
||||||
You get the best of speed and flexibility for your crazy research.
|
You get the best of speed and flexibility for your crazy research.
|
||||||
|
|
||||||
<p align=center><img width="80%" src="docs/source/_static/img/dynamic_graph.gif" /></p>
|

|
||||||
|
|
||||||
### Python First
|
### Python First
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue
Block a user