Commit Graph

17646 Commits

Author SHA1 Message Date
peter
3803d1c901 Fix conda build for Windows (#19824)
Summary:
Let's test it before merging.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19824

Differential Revision: D15116111

Pulled By: soumith

fbshipit-source-id: 0a73de3f045ee1349061674f5f8e2aaba382493c
2019-04-27 23:10:46 -07:00
Pieter Noordhuis
9b69da2b55 Allow for iterations where no module parameter is used (#19821)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19821

It is possible that not a single parameter is used during an
iteration. If this is the case, the `prepare_for_backward` function
marks all parameters as unused, kicks off reduction of all buckets,
*and* finalizes the reduction.

This is different from the prior implementation where we assumed that
autograd would produce a gradient for at least a single parameter.
We then used the autograd callback mechanism to queue a finalizer
callback. Now, this finalizer may be executed in line.

Reviewed By: mrshenli

Differential Revision: D15113272

fbshipit-source-id: dc91458b569cd8c106ddaeea558464b515683550
2019-04-27 22:57:59 -07:00
Michael Suo
f0a007a26c Use QualifiedName for classes (#19575)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19575
ghimport-source-id: eaed1d9d66672aadfe893e62a3a811da8ac7d966

Differential Revision: D15035281

Reviewed By: shannonzhu

Pulled By: suo

fbshipit-source-id: 7bac5e5f9223af77268bc03e08b37450a6840dbe
2019-04-27 16:13:27 -07:00
Michael Suo
1d6e868c2f make QualifiedName a value type (#19626)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19626
ghimport-source-id: 20f03b245245a7fab1c175cebc41e477b1eadc68

Reviewed By: shannonzhu

Differential Revision: D15049582

Pulled By: suo

fbshipit-source-id: 9e73a1d8ee8aa1849880815fa6fddebca408b8b4
2019-04-27 16:13:24 -07:00
Michael Suo
096dd8a4f2 separate QualifiedName into its own file (#19566)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19566
ghimport-source-id: c237f2a25d1aa9fc41f19fefe7a08a53a54279db

Differential Revision: D15032205

Reviewed By: shannonzhu

Pulled By: suo

fbshipit-source-id: 7527d97565559ebfb2556600eea5d93c1e141ac8
2019-04-27 16:13:20 -07:00
Chandler Zuo
472be69a73 Avoid Output Uninitialized Blobs in Load with load_all=1 (#19133)
Summary:
When output blob names are specified while load_all=1, output blob names are ignored. However, this behavior is not documented. In this diff, we just disallow users to provide blob names when load_all=1.

See discussion at https://fb.workplace.com/groups/1405155842844877/permalink/2714909788536136/
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19133

Reviewed By: dzhulgakov

Differential Revision: D14883698

Pulled By: chandlerzuo

fbshipit-source-id: 6e4171e36c4ccc4f857e79da98b858a06b7d8ad6
2019-04-27 10:45:44 -07:00
Max Wang
268859ce0d Fix CUDA stream syncing bug in allgather and reduce_scatter (#19631)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19631
ghimport-source-id: edc47e77d6ef03e966944ff98eefc22f2574eeaa

Reviewed By: mrshenli

Differential Revision: D15110077

Pulled By: mxw

fbshipit-source-id: 27a68308ade5ea511e2ea568a071eedb5d21c1ba
2019-04-27 08:35:56 -07:00
Michael Suo
a25b79531c use fully qualified name for ScriptClasses (#19239)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19239
ghimport-source-id: 830aad6dc11d2a7247760a9c7c9fc8556f70a706

Differential Revision: D14928293

Reviewed By: eellison

Pulled By: suo

fbshipit-source-id: d2efa5d7f7397526083278d6650b9cee8d967b1a
2019-04-26 19:17:21 -07:00
Xiaomeng Yang
2ce39de3fc Add elementwise_affine for layer_norm_op (#19713)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19713

Add elementwise_affine for layer_norm_op

Reviewed By: houseroad

Differential Revision: D15075454

fbshipit-source-id: e8a7d3da1c81e49fa55323f5e74a68bc4ef8d83f
2019-04-26 17:20:01 -07:00
David Riazati
f9786ad351 Add support for LONG_BINGET pickler op (#19815)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19815
ghimport-source-id: dd51c13892a8f0d91d726ae8ec65206d5e81f33e

Differential Revision: D15109969

Pulled By: driazati

fbshipit-source-id: da0bb5e30038173e74ca3e0e103dc11ba1638797
2019-04-26 17:13:48 -07:00
Elias Ellison
5a83a7424d fix optional type unification (#19813)
Summary:
Previously in type unification when we encountered an Optional[T] and a None, we would unify it to Optional[Optional[T]]. If you think about Optionals as a union of [T, None], then a union of [Optional[T], None] -> [T, None]. We should just be never create an Optional of an Optional.

The other fix is to change unify_types directly, but I think this is the more general fix, and would play more nicely with our optional type refinement, which also assumes we never encounter an Optional[Optional[T]].
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19813

Reviewed By: suo

Differential Revision: D15103083

Pulled By: eellison

fbshipit-source-id: db803db10d6934eaa5458e7c1746546b0d0c0a6c
2019-04-26 16:14:51 -07:00
Michael Antonov
698103cdd6 DataLoader docs update to describe how workers are managed, including Windows. (#18091)
Summary:
It's been hard to understand how workers are launched and what code runs in the worker vs. main process, especially on Windows, which leads to many of our samples failing. This explains when workers run an how to make code work on Windows as well.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/18091

Differential Revision: D15083766

Pulled By: soumith

fbshipit-source-id: 8a7e60defc8a72ec63874f657d7d5267d951dccf
2019-04-26 16:01:30 -07:00
Soumith Chintala
4e6608e86d Revert D15103223: [pytorch][PR] [CUDA 10] Resolve host_define.h warnings
Differential Revision:
D15103223

Original commit changeset: 5b56c4dd9cc4

fbshipit-source-id: f9a8e5ff0ee54cf5bb588896ab26dd9f0fb9ba45
2019-04-26 16:01:27 -07:00
Tongzhou Wang
42fbeef5d7 update F.grid_sample doc for clarity (#19754)
Summary:
https://github.com/pytorch/pytorch/issues/19717
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19754

Differential Revision: D15085449

Pulled By: soumith

fbshipit-source-id: 0dda05bd395d58a496bf397ca7f1c50a239b0ed1
2019-04-26 16:01:24 -07:00
davidriazati
dc67d9f3b9 Cleanup documentation (#19584)
Summary:
Stack from [ghstack](https://github.com/ezyang/ghstack):
* **#19584 [jit] Cleanup documentation**

Pull Request resolved: https://github.com/pytorch/pytorch/pull/19584

Pulled By: driazati

Differential Revision: D15104801

fbshipit-source-id: 87391fd62ee92b615e680469f8bd9a1ac654be7e
2019-04-26 15:43:07 -07:00
Soumith Chintala
75754beca3 Revert D14577575: [pytorch][PR] Fix lack of state init for adagrad and add share_memory flag
Differential Revision:
D14577575

Original commit changeset: 12440079ac96

fbshipit-source-id: 935106385e608471dc280fc61cfedf19d330812d
2019-04-26 15:43:04 -07:00
Orion Reblitz-Richardson
11297702b9 Fix the install of TensorBoard for doc generation (#19814)
Summary:
One more fix for https://github.com/pytorch/pytorch/pull/19810

We now know that we are running with python3, so no need to check python version. The quotes were probably causing problems here.

cc ezyang soumith zou3519
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19814

Differential Revision: D15106459

Pulled By: orionr

fbshipit-source-id: 0443b9b54d17fead9c8c2c9d8d2f373e1f95a28b
2019-04-26 14:56:04 -07:00
Stefan Krah
be20d65b70 Follow up to adaptive_max_pool3d() port (#19748)
Summary:
This is a follow up PR for #19547.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19748

Differential Revision: D15103230

Pulled By: ezyang

fbshipit-source-id: e7ce925faeadea502f77ed42d52e247c8c6571d8
2019-04-26 14:34:54 -07:00
Stefan Krah
cb4d41afcd Follow up to adaptive_max_pool2d() port (#19738)
Summary:
This is a follow up PR for #19409.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19738

Differential Revision: D15103231

Pulled By: ezyang

fbshipit-source-id: 11c9fec641b389906b8accd22504a683331fa6ec
2019-04-26 14:30:09 -07:00
Syed Tousif Ahmed
2573e695b0 Resolve host_define.h warnings (#19789)
Summary:
Eigen was updated with the commit needed to get rid of this warning that plagued the CI. This PR bumps third_party/eigen to that commit head.
```
warning: #warning "host_defines.h is an internal header file and must not be used directly.  This file will be removed in a future CUDA release.  Please use cuda_runtime_api.h or cuda_runtime.h instead." [-Wcpp]
```
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19789

Differential Revision: D15103223

Pulled By: ezyang

fbshipit-source-id: 5b56c4dd9cc41ff1794570ba2f6abfbe23f6ab68
2019-04-26 13:52:21 -07:00
Max Wang
c5845c4482 Add support for reduce-scatter in c10d (#18844)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/18844
ghimport-source-id: c6b2f0032c7c2212be2000a9c1f262f63d878a97

Stack from [ghstack](https://github.com/ezyang/ghstack):
* **#18844 Add support for reduce-scatter in c10d**
* #18820 Refactor ProcessGroupNCCL collective primitives

Reviewed By: mrshenli

Differential Revision: D14768369

fbshipit-source-id: a9def7a0da6e9cd995e982371cc1e22f3df1a156
2019-04-26 13:46:57 -07:00
Junjie Bai
c9f380df02 Add aten mkldnn linear operator
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19210

Reviewed By: dzhulgakov

Differential Revision: D14901641

fbshipit-source-id: 8fa68b9941fd93cea0f313a828cba34c5c81ae11
2019-04-26 13:41:57 -07:00
Junjie Bai
48b81da4cb Add aten mkldnn view operator
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19209

Reviewed By: dzhulgakov

Differential Revision: D14894545

fbshipit-source-id: 69455184811de1d1444b5d494e4a9d8c83301431
2019-04-26 13:41:54 -07:00
Junjie Bai
61d5a8dded Add aten mkldnn add operator
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19207

Reviewed By: dzhulgakov

Differential Revision: D14889477

fbshipit-source-id: 2c5e5ea5dfc26a9c9a172c5fa2c6d7584b167e16
2019-04-26 13:41:51 -07:00
Junjie Bai
fb53c189b3 Add aten mkldnn batch_norm operator
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19206

Reviewed By: dzhulgakov

Differential Revision: D14887205

fbshipit-source-id: ea00c9e3205c449d08ab29535309164f951aab95
2019-04-26 13:41:48 -07:00
Junjie Bai
4864000e55 Add aten mkldnn ops: relu, max_pool2d and avg_pool2d
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19205

Reviewed By: dzhulgakov

Differential Revision: D14850598

fbshipit-source-id: 5bbd5909c06df9c980de680ffb81bf772766c0ba
2019-04-26 13:41:44 -07:00
Junjie Bai
3445020ca3 Add aten mkldnn conv2d operator
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19204

Reviewed By: dzhulgakov

Differential Revision: D14857513

fbshipit-source-id: 1172c9785e5a17a7d7360474551bdc7a511b3f2f
2019-04-26 13:41:41 -07:00
Junjie Bai
8f1445c406 Add is_mkldnn to at::Tensor
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19633

Reviewed By: dzhulgakov

Differential Revision: D15053320

fbshipit-source-id: 12b9f85a025a9e957e1b7b3014ba44ae71bfd7a5
2019-04-26 13:41:38 -07:00
Wanchao Liang
236c2b2387 Let script module buffer attributes can also cast device/type (#19700)
Summary:
Tested locally this  fix #19039, did not add a test since there's no way to create a script module in the cpp world.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19700

Differential Revision: D15094195

Pulled By: wanchaol

fbshipit-source-id: fcc2c1e5efbc160d976ae485ba2457442f62f065
2019-04-26 13:06:52 -07:00
Will Feng
5099db08d4 Ignore nn::Functional submodules in nn::Module serialization (#19740)
Summary:
Currently, the Python API doesn't serialize layers that don't have weights (such as `nn.ReLU` and `nn.MaxPool2d`e.g. in https://github.com/pytorch/vision/blob/master/torchvision/models/densenet.py#L80-L81). If one saves a model that contains weight-less layers in Python and tries to load it into C++, the C++ module loading code (`torch::load(...)`) will throw an error complaining that the expected layers are not found in the serialized file (e.g. https://github.com/pytorch/vision/pull/728#issuecomment-480974175). This PR solves the problem by ignoring layers that are not serializable (which currently only include `nn::Functional`) in the C++ module serialization code (`torch::save(...)` and `torch::load(...)`), and the user is expected to use `nn::Functional` to wrap the weight-less layers so that they can be ignored when serializing / deserializing.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19740

Differential Revision: D15100575

Pulled By: yf225

fbshipit-source-id: 956481a2355d1de45341585abedda05e35d2ee8b
2019-04-26 12:47:23 -07:00
Max Wang
61d48aa989 Refactor ProcessGroupNCCL collective primitives (#18820)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/18820
ghimport-source-id: 220b2a3dd9d4d6d2e557e1802851f082c2dc6452

Stack from [ghstack](https://github.com/ezyang/ghstack):
* **#18820 Refactor ProcessGroupNCCL collective primitives**

Planning to add reduce-scatter, but no room in my stomach for more
copypasta.

Also rewrote the tensor list validation logic.  The existing validation
was ill-suited for all the cases it was being used for; it took a vector
of input tensors and a vector of output tensors, but only ever received
either two references to the same vector, or a bespoke singleton vector
and a vector of outputs (for which it would ignore all but the first
output).  In the first case, it performed unnecessary checks, and in the
second, it skipped necessary ones.

Reviewed By: mrshenli

Differential Revision: D14762369

fbshipit-source-id: dcf882ce1c5854333a9eb4424bfc18d9f4648ddf
2019-04-26 12:38:48 -07:00
Orion Reblitz-Richardson
e1ebf330d5 Install TensorBoard for doc generation (#19810)
Summary:
In order to have `torch.utils.tensorboard.SummaryWriter` rendered in the documentation at the bottom of https://pytorch.org/docs/master/tensorboard.html we need to have TensorBoard installed.

This change makes it so our pinned version of `tb-nightly` is used for doc generation same as it is used for running tests at https://github.com/pytorch/pytorch/blob/master/.jenkins/pytorch/test.sh#L45-L52

Eventually we'll use a pinned version of `pip install tensorboard`, but it's not on the release channel yet.

cc kostmo soumith ezyang
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19810

Differential Revision: D15101730

Pulled By: orionr

fbshipit-source-id: c41678c4f9ef3d56a168f2b96a1ab05f351bdc56
2019-04-26 12:06:18 -07:00
huangyanhua
bacc8815c7 update Anaconda download link (#19794)
Summary:
Now `https://www.continuum.io/` is redirected to `https://www.anaconda.com` and old Anaconda download link `https://www.continuum.io/downloads` is dead. This PR update it to `https://www.anaconda.com/distribution/#download-section`.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19794

Differential Revision: D15099538

Pulled By: soumith

fbshipit-source-id: 967dcda34d9d446c0d26c0014f10cc710f69a0c5
2019-04-26 09:45:44 -07:00
Spandan Tiwari
dafee117e8 Removing unused arg f from _model_to_graph(). (#19647)
Summary:
Input argument `f` in `_model_to_graph()` method in `torch/onnx/utils.py` is unused. This PR removes it. If there's a reason to keep it around, please let me know.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19647

Reviewed By: dzhulgakov

Differential Revision: D15071720

Pulled By: houseroad

fbshipit-source-id: 59e0dd7a4d5ebd64d0e30f274b3892a4d218c496
2019-04-26 09:40:52 -07:00
Pieter Noordhuis
0d8a3610c5 Multiple module outputs and multiple calls to backward (#19799)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19799

A module that returns multiple outputs and where the called may end up
doing multiple calls to torch.autograd.backward did not work with
DistributedDataParallel. It expected the first call to
torch.autograd.backward to provide gradients for ALL parameters that
expect gradients and were used in computing the module output. If you
have outputs with disjoint autograd graphs it is fine to call
torch.autograd.backward on both and fill in the module's parameter
gradients in separate chunks.

With this change we delay queuing the finalizer callback until we have
marked all buckets as ready, instead of queueing it the first time we
receive an autograd hook. This returns the current implementation to
be functionally equivalent to the DistributedDataParallel
implementation before #18953 was merged.

Reviewed By: mrshenli

Differential Revision: D15097045

fbshipit-source-id: 2df023319713bc31e29a8b45108c78e6593fccd4
2019-04-26 08:20:10 -07:00
Eric Faust
dcfb5620df Allow passing lists as trace inputs.
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19580

Differential Revision: D15034978

fbshipit-source-id: d3bc32ccae1c12104f2bde43fd4700d220bb3ca9
2019-04-26 02:41:57 -07:00
Karl Ostmo
8f0603b128 C++ changes toward libtorch and libcaffe2 unification (#19554)
Summary:
* adds TORCH_API and AT_CUDA_API in places
* refactor code generation Python logic to separate
  caffe2/torch outputs
* fix hip and asan
* remove profiler_cuda from hip
* fix gcc warnings for enums
* Fix PythonOp::Kind
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19554

Differential Revision: D15082727

Pulled By: kostmo

fbshipit-source-id: 83a8a99717f025ab44b29608848928d76b3147a4
2019-04-26 01:38:10 -07:00
Yinghai Lu
9d180e602f More topi support (#19728)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19728

Added `Tanh`, `Transpose` and `Mul` support.

Reviewed By: hlu1

Differential Revision: D15078878

fbshipit-source-id: 0a0df6b0d453bc38987b6d744774c127dd6875fe
2019-04-26 00:53:11 -07:00
zrphercule
c182824f69 Update foxi version (#19793)
Summary:
Update foxi to the latest version for group quantization.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19793

Reviewed By: jackm321, houseroad

Differential Revision: D15095982

Pulled By: zrphercule

fbshipit-source-id: 0d1cb403cbda47a4fda9035e1712fced60ced283
2019-04-25 22:39:40 -07:00
Lu Fang
20c22bcae4 Automatic update of fbcode/onnx to 22662bfd4dcc6baebf29e3b823a051676f991001 (#19790)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19790

Previous import was 27d4b617e7097cda7d0d4c45ff2b09d248f33179

Included changes:
- **[22662bfd](https://github.com/onnx/onnx/commit/22662bfd)**: Bump up version number and update Versioning for 1.5.0 release (#1965) <Raymond Yang>
- **[b1a3a8c8](https://github.com/onnx/onnx/commit/b1a3a8c8)**: fix the ci (#1964) <Lu Fang>

Reviewed By: zrphercule

Differential Revision: D15095183

fbshipit-source-id: b69cb62685122b83a1493b2702aa6ec950ee15bf
2019-04-25 22:23:25 -07:00
Junjie Bai
f0d493d290 Add devtoolset 8 (gcc 8) + glibc 2.26 + centos 7.5 rocm docker image (#19767)
Summary:
xw285cornell

Will add py3.6-devtoolset8-glibc2.26-rocmrpm-centos7.5
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19767

Differential Revision: D15094446

Pulled By: bddppq

fbshipit-source-id: 01a932d893cf4559f98612888308b3ad6900a038
2019-04-25 22:13:20 -07:00
Tzu-Wei Huang
98e312cf96 TensorBoard support within PyTorch (#16196)
Summary:
This PR adds TensorBoard logging support natively within PyTorch. It is based on the tensorboardX  code developed by lanpa and relies on changes inside the tensorflow/tensorboard repo landing at https://github.com/tensorflow/tensorboard/pull/2065.

With  these changes users can simply `pip install tensorboard; pip install torch` and then log PyTorch data directly to the TensorBoard protobuf format using

```
import torch
from torch.utils.tensorboard import SummaryWriter
writer = SummaryWriter()
s1 = torch.rand(1)
writer.add_scalar('data/scalar1', s1[0], 0)
writer.close()
```

Design:
- `EventFileWriter` and `RecordWriter` from tensorboardX now live in tensorflow/tensorboard
- `SummaryWriter` and PyTorch-specific conversion from tensors, nn modules, etc. now live in pytorch/pytorch. We also support Caffe2 blobs and nets.

Action items:
- [x] `from torch.utils.tensorboard import SummaryWriter`
- [x] rename functions
- [x] unittests
- [x] move actual writing function to tensorflow/tensorboard in https://github.com/tensorflow/tensorboard/pull/2065

Review:
- Please review for PyTorch standard formatting, code usage, etc.
- Please verify unittest usage is correct and executing in CI

Any significant changes made here will likely be synced back to github.com/lanpa/tensorboardX/ in the future.

cc orionr, ezyang
Pull Request resolved: https://github.com/pytorch/pytorch/pull/16196

Differential Revision: D15062901

Pulled By: orionr

fbshipit-source-id: 3812eb6aa07a2811979c5c7b70810261f9ea169e
2019-04-25 21:30:23 -07:00
Ailing Zhang
97e80ab6fc Always enable autodiff check (#19787)
Summary:
disable_autodiff_subgraph_inlining should be always on to check AD regression.
Thanks eellison for spotting the test regression!
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19787

Differential Revision: D15093104

Pulled By: ailzhang

fbshipit-source-id: 82a75a7dd7097d5f93a2e4074023da2105341c1b
2019-04-25 21:22:30 -07:00
Jack Montgomery
48d5ab54a8 Automatic update of fbcode/foxi to 8f74bc4df3a4cfc69b1a3eadf62aa29d9961c72d AND update Glow AND update C2 (#19792)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19792

This diff also contains the contents of D15092641 and D15090411 so as to not let c2, foxi, and glow get out of sync

Previous import was 81e1683d6348eee4b5ed1145222dc2c41be4269c

Included changes:
- **[8f74bc4](https://github.com/houseroad/foxi/commit/8f74bc4)**: Small fixes (#12) <Jack Montgomery>
- **[72097e4](https://github.com/houseroad/foxi/commit/72097e4)**: Add multiple quantization params per tensor (#11) <Jack Montgomery>
- **[b681fe0](https://github.com/houseroad/foxi/commit/b681fe0)**: Merge pull request #10 from jackm321/add_autoinstrument_graph_prop <Jack Montgomery>
- **[a68d835](https://github.com/houseroad/foxi/commit/a68d835)**: Add ONNXIFI_GRAPH_PROPERTY_AUTO_INSTRUMENT_NODES <Jack Montgomery>

Reviewed By: rdzhabarov, zrphercule

Differential Revision: D15086794

fbshipit-source-id: 8df02c62303b580e16a218d6be7791747e3d7213
2019-04-25 21:03:32 -07:00
Alexander Sidorov
7a8bc85f47 Profiler: add Self CPU Time Total, CPU time total and other general improvements (#19378)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19378

Function profile events are typically nested. In this diff I
add parent child relationship to the intervals. This way we can
attribute self time easily. As a result, user printing a table from a
profiler trace gets self cpu time.

This diff doesn't try to address CUDA self time as CUDA kernels are
already getting special care in the profiler.

There are also some other minor improvements. Like reporting total CPU
time spent, reversed sorting, aggregated data after the table,
etc.

There is a new unit test added which tests more functionality than
previous profiler test

Reviewed By: zheng-xq

Differential Revision: D14988612

fbshipit-source-id: 2ee6f64f0a4d0b659c6b23c0510bf13aa46f07dc
2019-04-25 20:53:55 -07:00
Zafar Takhirov
6e06154c13 Quantized SumRelu (#19319)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19319

Quantized SUM + ReLU (Fused). The implementation is the same as the one in the DNNLOWP.

Reviewed By: jianyuh

Differential Revision: D14866442

fbshipit-source-id: c8c737a37e35b6ce3c1c2077c07546aba16e0612
2019-04-25 18:01:21 -07:00
Zafar Takhirov
76307667ca Use the QTensor with QReLU (#19312)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19312

Replaces the tuple hack with the QTensor. Please, note this can be landed ONLY after #18960 (D14810261) is landed.

Reviewed By: raghuramank100

Differential Revision: D14819460

fbshipit-source-id: 75ca649304b1619cb3cfe845962c9f226b8f884a
2019-04-25 18:01:17 -07:00
Zafar Takhirov
db9008496e Changing the rounding in the QTensor (#19714)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/19714

We had rounding in the quantizer set as `round(x/scale) + zp`. To make it consistent, converting it to `round(x/scale + zp)`.

Reviewed By: raghuramank100

Differential Revision: D15077095

fbshipit-source-id: 5d20a90391fe8c2e11b338c05631fcf7770320c3
2019-04-25 18:01:13 -07:00
Jesse Hellemn
e814c11045 Fix env vars needed for devtoolset7 binaries
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19780

Differential Revision: D15091963

Pulled By: pjh5

fbshipit-source-id: 2594395b2313d5c8a37db28965d99b0541a227e3
2019-04-25 17:50:14 -07:00
Jesse Hellemn
c5cca65351 Fixing update_s3_htmls for binaries
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/19746

Differential Revision: D15091326

Pulled By: pjh5

fbshipit-source-id: ed172c678dd5659fa31d5d9b6ee1bf119ede2889
2019-04-25 17:24:02 -07:00