Commit Graph

17 Commits

Author SHA1 Message Date
Simon Layton
84e7eff458 Waive some hypothesis tests on GPU
Summary:
operators don't exist on GPU
Closes https://github.com/caffe2/caffe2/pull/63

Reviewed By: Yangqing

Differential Revision: D4348968

Pulled By: bwasti

fbshipit-source-id: 1fb8693842d6827ffcf96de2a9a8ba2f9dff0293
2016-12-19 15:59:32 -08:00
Yangqing Jia
42bbdda8c4 MKLDevice and MKLOperator
Summary:
This adds Caffe2 support for MKL operators directly with MKLMemory. Included a
Relu layer that shows how to use it.

Reviewed By: salexspb

Differential Revision: D4322144

fbshipit-source-id: 8b3392c4fd024ab1a7ba7135c349ebd3e1976799
2016-12-15 19:59:24 -08:00
Yangqing Jia
dc16bcfa27 Remove float64 test
Summary:
float64 test breaks things on the cuda side. I am deleting it for now and if
we add it back, let's make sure we run the test on a GPU machine first :)

Reviewed By: azzolini

Differential Revision: D4324427

fbshipit-source-id: 0246fe9dd28a286422ca94c90f5b0fc33a162e74
2016-12-15 12:01:30 -08:00
Maxime Boucher
4cd263db74 Last N window collector
Summary: Allows to collect samples over multiple batches. The method uses a circular array and so there is no guarantee about the order of the samples. The goal is to get a view of the data accross multiple batches

Reviewed By: salexspb

Differential Revision: D4216181

fbshipit-source-id: bb9e1fa84ac7e04006dcddb53c9347a42ec83dc8
2016-12-15 12:01:30 -08:00
Xianjie Chen
0bc104a3d0 fix unit test
Summary: ...

Differential Revision: D4298663

fbshipit-source-id: 7831830a5201eb6603d846460c22b2f906e53858
2016-12-15 12:01:29 -08:00
Xianjie Chen
3c47d41f86 add unit test for row mul
Summary: so that we are more confident.

Differential Revision: D4290132

fbshipit-source-id: 44e4687d977ab90cc022a14131bbf701bdf131d4
2016-12-15 12:01:29 -08:00
Xianjie Chen
f41b2ca85c fix sliceop for empty batch
Summary: Used in the NNPreProc layers. It fails the online training when there is empty batch.

Reviewed By: dzhulgakov

Differential Revision: D4235498

fbshipit-source-id: bde00a011831762e44a3f9bf2190d4b241a06ccc
2016-11-29 15:18:39 -08:00
Wenlin Chen
9fa26fcc32 position weighted embedding
Summary: Each sparse feature is a ID list. And usually the position of the id in the id list is meaningful. The earlier the id appears in the list, the more important. In this diff, we multiple each embedding with a weight, where the weight corresponds to the position. With this change, same ID appears on different position would have different norm/length/importance after aggregation. The firstX transformation in sigrid is a special case of this model where the weights before n are 1, and 0 after n, where n is the argument of firstX.

Reviewed By: xianjiec

Differential Revision: D4181251

fbshipit-source-id: 2a6f8b7240af445b6bd2052fd24c2d99f39ee7ff
2016-11-29 15:18:35 -08:00
Yangqing Jia
589398950f fbsync at f5a877 2016-11-18 15:41:06 -08:00
Yangqing Jia
238ceab825 fbsync. TODO: check if build files need update. 2016-11-15 00:00:46 -08:00
Yangqing Jia
d1e9215184 fbsync 2016-10-07 13:08:53 -07:00
Yangqing Jia
b23e51d467 chunky sync 2016-09-06 15:55:19 -07:00
Yangqing Jia
05512d1e10 sync 2016-08-10 11:02:15 -07:00
Yangqing Jia
c15e45c9bb chunky sync again 2016-08-01 20:58:46 -07:00
Yangqing Jia
bcea409c82 sync 2016-07-28 15:06:43 -07:00
Yangqing Jia
6463eebc7b chunky sync - build scripts to be written 2016-07-21 10:16:42 -07:00
Yangqing Jia
559053d3a8 chunky sync 2016-05-13 14:43:48 -07:00