..
experimental/c10 /cpu
Delete C10Tensor ( #19328 )
2019-04-17 00:02:27 -07:00
hip
Fix caffe2 miopen conv transpose gradient op for case of no dX gradient
2019-04-04 17:29:30 -07:00
quantized
Adding quantized tensor shape/type info support for caffe2=>glow in caffe2 side ( #18621 )
2019-03-31 17:42:27 -07:00
rnn
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
abs_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
abs_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
abs_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
accumulate_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
accumulate_op.cu
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
accumulate_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
accuracy_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
accuracy_op.cu
Tensor method rename ndim()->dim() - 1/3 ( #16678 )
2019-02-04 15:49:16 -08:00
accuracy_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
acos_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
acos_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
acos_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
activation_ops_cudnn.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
affine_channel_op.cc
Tensor construction codemod(ResizeLike) - 3/7 ( #15122 )
2018-12-14 02:08:37 -08:00
affine_channel_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
affine_channel_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
apmeter_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
apmeter_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
arg_ops.cc
Remove many caffe2::TIndex and replace them with int64_t ( #11943 )
2018-09-22 18:11:04 -07:00
arg_ops.cu
Remove many caffe2::TIndex and replace them with int64_t ( #11943 )
2018-09-22 18:11:04 -07:00
arg_ops.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
asin_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
asin_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
asin_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
assert_op.cc
Convert all tabs to spaces, add CI. ( #18959 )
2019-04-09 08:12:26 -07:00
assert_op.cu
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
assert_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
atan_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
atan_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
atan_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
atomic_ops.cc
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
batch_box_cox_op.cc
Tensor construction codemod(ResizeLike) - 3/7 ( #15122 )
2018-12-14 02:08:37 -08:00
batch_box_cox_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
batch_bucketize_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
batch_bucketize_op.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
batch_gather_ops.cc
Implements Gather operator for arbitrary axis, sharing the code with BatchGather. ( #13756 )
2018-12-04 11:54:28 -08:00
batch_gather_ops.cu
Fix type-o in unsupported data type error message ( #16537 )
2019-02-05 10:46:46 -08:00
batch_gather_ops.h
refactor caffe2 operator constructors - 1/9 ( #17082 )
2019-03-04 16:04:01 -08:00
batch_matmul_op_gpu_test.cc
Renaming size() to numel() - 1/6
2018-10-29 11:11:19 -07:00
batch_matmul_op_test.cc
Renaming size() to numel() - 1/6
2018-10-29 11:11:19 -07:00
batch_matmul_op.cc
optimize BatchMatmulOp ( #18612 )
2019-04-23 15:34:59 -07:00
batch_matmul_op.cu
optimize BatchMatmulOp ( #18612 )
2019-04-23 15:34:59 -07:00
batch_matmul_op.h
optimize BatchMatmulOp ( #18612 )
2019-04-23 15:34:59 -07:00
batch_moments_op.cc
Optimize reduce ops for 2d and 3d ( #9992 )
2018-08-04 13:53:58 -07:00
batch_moments_op.cu
Add implementation for batch_moments_op ( #9510 )
2018-07-18 11:59:54 -07:00
batch_moments_op.h
refactor caffe2 operator constructors - 10/9 ( #17659 )
2019-03-06 15:11:47 -08:00
batch_sparse_to_dense_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
batch_sparse_to_dense_op.h
refactor caffe2 operator constructors - 10/9 ( #17659 )
2019-03-06 15:11:47 -08:00
bbox_transform_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
bbox_transform_op.h
Improve caffe2 operator wrapping ( #17743 )
2019-03-14 08:54:15 -07:00
bisect_percentile_op.cc
Add a bisect percentile operator ( #10563 )
2018-08-20 13:14:05 -07:00
bisect_percentile_op.h
refactor caffe2 operator constructors - 10/9 ( #17659 )
2019-03-06 15:11:47 -08:00
boolean_mask_ops.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
boolean_mask_ops.cu
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
boolean_mask_ops.h
refactor caffe2 operator constructors - 10/9 ( #17659 )
2019-03-06 15:11:47 -08:00
boolean_unmask_ops_test.cc
Renaming size() to numel() - 1/6
2018-10-29 11:11:19 -07:00
boolean_unmask_ops.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
boolean_unmask_ops.cu
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
boolean_unmask_ops.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
box_with_nms_limit_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
box_with_nms_limit_op.h
register BoxWithNMSLimit with C10
2019-03-29 13:41:40 -07:00
byte_weight_dequant_op.cc
Add byte_weight_dequant_op
2018-07-18 16:27:21 -07:00
byte_weight_dequant_op.h
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
cast_op.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
cast_op.cu
Tensor construction codemod ( #16568 )
2019-02-05 18:51:02 -08:00
cast_op.h
refactor caffe2 operator constructors - 11/9 ( #17722 )
2019-03-08 12:38:54 -08:00
cbrt_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
cbrt_op.cu
Add Cube and Cbrt Ops ( #8991 )
2018-06-28 14:55:30 -07:00
cbrt_op.h
Use FixedDivisor in Reduce and Broadcast CUDA kernels ( #9072 )
2018-07-01 00:25:34 -07:00
cc_bmm_bg_op.cc
Move ConcatBatchMatMulBatchGatherOp to OSS
2019-04-10 15:29:03 -07:00
cc_bmm_bg_op.h
Move ConcatBatchMatMulBatchGatherOp to OSS
2019-04-10 15:29:03 -07:00
ceil_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
ceil_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
ceil_op.h
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
channel_backprop_stats_op.cc
Rename ndim() -> dim() - 3/6
2018-11-05 23:21:40 -08:00
channel_backprop_stats_op.cu
Tensor method rename ndim()->dim() - 1/3 ( #16678 )
2019-02-04 15:49:16 -08:00
channel_backprop_stats_op.h
refactor caffe2 operator constructors - 11/9 ( #17722 )
2019-03-08 12:38:54 -08:00
channel_shuffle_op.cc
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
channel_shuffle_op.cu
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
channel_shuffle_op.h
refactor caffe2 operator constructors - 11/9 ( #17722 )
2019-03-08 12:38:54 -08:00
channel_stats_op.cc
Optimize channel_stats_op ( #16243 )
2019-03-12 12:08:00 -07:00
channel_stats_op.cu
Optimize channel_stats_op ( #16243 )
2019-03-12 12:08:00 -07:00
channel_stats_op.h
Optimize channel_stats_op ( #16243 )
2019-03-12 12:08:00 -07:00
clip_op.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
clip_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
clip_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
CMakeLists.txt
Add functor- and function-based kernel registration API ( #18162 )
2019-03-30 00:07:19 -07:00
collect_and_distribute_fpn_rpn_proposals_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
collect_and_distribute_fpn_rpn_proposals_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
communicator_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
communicator_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
concat_split_op_gpu.cc
Update from facebook ( #7451 )
2018-05-10 23:14:27 -07:00
concat_split_op.cc
Add shape inference function for Split ( #18838 )
2019-04-04 00:22:22 -07:00
concat_split_op.h
Add backward pass to infer single missing input shape for Concat opportunitiscally ( #18911 )
2019-04-05 10:11:58 -07:00
conditional_op.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
conditional_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
conv_gradient_op.cc
Fix conv gradient conversion ( #11312 )
2018-09-05 20:09:31 -07:00
conv_op_cache_cudnn_test.cc
move flags to c10 ( #12144 )
2018-10-04 02:09:56 -07:00
conv_op_cache_cudnn.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
conv_op_cache_cudnn.h
Rename IntList to IntArrayRef. ( #16751 )
2019-02-05 14:54:34 -08:00
conv_op_cudnn.cc
Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize ( #17764 )
2019-03-07 18:38:53 -08:00
conv_op_eigen.cc
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
conv_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
conv_op_impl.h
fix -Wsign-compare warnings for some files inside c2 ( #18123 )
2019-03-19 10:39:20 -07:00
conv_op_shared_gpu.cc
Removing some dependency edges from Blob to other caffe2 ( #12043 )
2018-09-25 11:40:24 -07:00
conv_op_shared.cc
move flags to c10 ( #12144 )
2018-10-04 02:09:56 -07:00
conv_op_shared.h
Remove template parameter from Tensor ( #9939 )
2018-07-27 10:56:39 -07:00
conv_op.cc
Convert some docstrings from char* to char[] ( #13062 )
2018-10-24 13:48:18 -07:00
conv_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
conv_pool_op_base.h
Add NHWC order support in the cost inference function of 3d conv ( #19170 )
2019-04-15 16:47:22 -07:00
conv_transpose_gradient_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
conv_transpose_op_cudnn.cc
Add support for group ConvTranspose ( #18794 )
2019-04-04 11:52:06 -07:00
conv_transpose_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
conv_transpose_op_impl.h
Add support for group ConvTranspose ( #18794 )
2019-04-04 11:52:06 -07:00
conv_transpose_op_mobile_impl.h
Make C10_MOBILE consistent with how feature macros are usually used ( #17481 )
2019-02-27 17:57:51 -08:00
conv_transpose_op_mobile_test.cc
Renaming size() to numel() - 1/6
2018-10-29 11:11:19 -07:00
conv_transpose_op_mobile.cc
Make C10_MOBILE consistent with how feature macros are usually used ( #17481 )
2019-02-27 17:57:51 -08:00
conv_transpose_op_mobile.h
Make C10_MOBILE consistent with how feature macros are usually used ( #17481 )
2019-02-27 17:57:51 -08:00
conv_transpose_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
conv_transpose_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
conv_transpose_unpool_op_base.h
Add support for group ConvTranspose ( #18794 )
2019-04-04 11:52:06 -07:00
copy_op.cc
Split out copy_op from utility_ops ( #11470 )
2018-09-12 16:25:48 -07:00
copy_op.cu
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
copy_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
cos_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
cos_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
cos_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
cosh_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
cosh_op.cu
Implement Sinh and Cosh ( #9213 )
2018-07-10 18:55:31 -07:00
cosh_op.h
Implement Sinh and Cosh ( #9213 )
2018-07-10 18:55:31 -07:00
cosine_embedding_criterion_op.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
cosine_embedding_criterion_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
cosine_embedding_criterion_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
counter_ops_gpu.cc
HIP Operators Generator--> HipOpG ( #9322 )
2018-07-19 00:26:06 -07:00
counter_ops.cc
Convert all tabs to spaces, add CI. ( #18959 )
2019-04-09 08:12:26 -07:00
counter_ops.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
crash_op.cc
Support compilation on gcc-7.4.0 ( #19470 )
2019-04-19 21:41:36 -07:00
create_scope_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
create_scope_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
crf_viterbi_op.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
cross_entropy_op.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
cross_entropy_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
cross_entropy_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
ctc_beam_search_decoder_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
ctc_beam_search_decoder_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
ctc_greedy_decoder_op.cc
fix -Wsign-compare warnings for some files inside c2 ( #18123 )
2019-03-19 10:39:20 -07:00
ctc_greedy_decoder_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
cube_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
cube_op.cu
Add Cube and Cbrt Ops ( #8991 )
2018-06-28 14:55:30 -07:00
cube_op.h
Add Cube and Cbrt Ops ( #8991 )
2018-06-28 14:55:30 -07:00
data_couple_gpu.cu
No Op Optimizer ( #12390 )
2018-10-10 18:09:46 -07:00
data_couple.cc
No Op Optimizer ( #12390 )
2018-10-10 18:09:46 -07:00
data_couple.h
No Op Optimizer ( #12390 )
2018-10-10 18:09:46 -07:00
dataset_ops.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
dataset_ops.h
Renaming meta() to dtype() - 1/2 ( #13333 )
2018-10-31 17:14:08 -07:00
deform_conv_gradient_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
deform_conv_op_impl.h
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
deform_conv_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
deform_conv_op.cu
Rename IntList to IntArrayRef. ( #16751 )
2019-02-05 14:54:34 -08:00
deform_conv_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
dense_vector_to_id_list_op.cc
add dense vector to id_list operator ( #15090 )
2018-12-18 16:27:38 -08:00
dense_vector_to_id_list_op.h
Tensor construction codemod ( #16568 )
2019-02-05 18:51:02 -08:00
depthwise_3x3_conv_op_cudnn.cu
Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize ( #17764 )
2019-03-07 18:38:53 -08:00
distance_op.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
distance_op.cu
Tensor construction codemod ( #16568 )
2019-02-05 18:51:02 -08:00
distance_op.h
Move math::Axpy function to elementwise lib ( #18316 )
2019-03-26 12:19:19 -07:00
do_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
do_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
do_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
dropout_op_cudnn.cc
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
dropout_op.cc
Tensor construction: combine Resize+mutable_data - 1/4 ( #13942 )
2018-11-19 15:33:50 -08:00
dropout_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
dropout_op.h
refactor caffe2 operator constructors - 2/9 ( #17083 )
2019-02-28 14:23:55 -08:00
elementwise_add_gradient_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_add_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
elementwise_add_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_add_op.h
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
elementwise_div_gradient_op.cc
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
elementwise_div_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_div_op.cu
Remaining Tensor API fixes - dims() -> sizes() ( #15743 )
2019-01-15 18:42:02 -08:00
elementwise_div_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
elementwise_linear_op.cc
Export ElementwiseLinear to ONNX (Mul + Add). ( #17411 )
2019-02-25 08:11:13 -08:00
elementwise_linear_op.cu
Tensor method rename ndim()->dim() - 2/3 ( #16679 )
2019-02-04 11:12:57 -08:00
elementwise_linear_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
elementwise_logical_ops.cc
Windows raw string fix ( #10998 )
2018-08-29 11:40:08 -07:00
elementwise_logical_ops.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
elementwise_mul_gradient_op.cc
Avoid undefined symbol error when building AdIndexer LTO ( #19009 )
2019-04-15 23:43:13 -07:00
elementwise_mul_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_mul_op.cu
Use FixedDivisor in Reduce and Broadcast CUDA kernels ( #9072 )
2018-07-01 00:25:34 -07:00
elementwise_mul_op.h
Avoid undefined symbol error when building AdIndexer LTO ( #19009 )
2019-04-15 23:43:13 -07:00
elementwise_op_gpu_test.cc
move flags to c10 ( #12144 )
2018-10-04 02:09:56 -07:00
elementwise_op_test.cc
move flags to c10 ( #12144 )
2018-10-04 02:09:56 -07:00
elementwise_op_test.h
Renaming size() to numel() - 2/6
2018-10-26 15:21:50 -07:00
elementwise_ops_schema.cc
Fix typo in Github links in elementwise_ops_schema.cc ( #18018 )
2019-03-26 15:37:26 -07:00
elementwise_ops_utils.cc
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
elementwise_ops_utils.h
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
elementwise_ops.cc
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
elementwise_ops.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
elementwise_ops.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
elementwise_sub_gradient_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_sub_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
elementwise_sub_op.cc
[caffe2] Move elementwise grad ops to separate files ( #8315 )
2018-06-11 15:38:36 -07:00
elementwise_sub_op.h
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
elementwise_sum_op.cc
add tensor and cost inference functions ( #17684 )
2019-03-06 23:34:02 -08:00
elu_op_cudnn.cc
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
elu_op.cc
Use REGISTER_CPU_GRADIENT_OPERATOR for many operators ( #12616 )
2018-10-24 13:48:17 -07:00
elu_op.cu
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
elu_op.h
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
enforce_finite_op.cc
Update from Facebook ( #6692 )
2018-04-17 23:36:40 -07:00
enforce_finite_op.cu
Remove calls to CopyFrom that can be sync ( #13205 )
2018-10-29 13:57:42 -07:00
enforce_finite_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
ensure_clipped_op.cc
Renaming size() to numel() - 2/6
2018-10-26 15:21:50 -07:00
ensure_clipped_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
ensure_cpu_output_op.cc
Update from facebook ( #7451 )
2018-05-10 23:14:27 -07:00
ensure_cpu_output_op.cu
Update from facebook ( #7451 )
2018-05-10 23:14:27 -07:00
ensure_cpu_output_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
erf_op.cc
Export PyTorch erf to ONNX Erf and add Caffe2 Erf operator
2019-01-17 09:18:08 -08:00
erf_op.cu
Export PyTorch erf to ONNX Erf and add Caffe2 Erf operator
2019-01-17 09:18:08 -08:00
erf_op.h
Export PyTorch erf to ONNX Erf and add Caffe2 Erf operator
2019-01-17 09:18:08 -08:00
exp_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
exp_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
exp_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
expand_op_gpu.cc
New operator 'expand' ( #8263 )
2018-06-18 16:33:47 -07:00
expand_op.cc
Convert all tabs to spaces, add CI. ( #18959 )
2019-04-09 08:12:26 -07:00
expand_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
expand_squeeze_dims_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
expand_squeeze_dims_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
expand_squeeze_dims_op.h
fix -Wsign-compare warnings for some files inside c2 ( #18123 )
2019-03-19 10:39:20 -07:00
fc_inference.cc
remove redundant inference functions for FC ( #17407 )
2019-02-22 20:13:20 -08:00
fc_inference.h
remove redundant inference functions for FC ( #17407 )
2019-02-22 20:13:20 -08:00
feature_maps_ops.cc
[Caffe2][fbcode=>GH sync] Update from facebook 4323b18ce13c ( #7116 )
2018-05-01 20:49:00 -07:00
feature_maps_ops.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
feed_blob_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
feed_blob_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
filler_op.cc
Renaming size() to numel() - 2/6
2018-10-26 15:21:50 -07:00
filler_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
filler_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
find_duplicate_elements_op.cc
Updates to caffe2 operator documentation ( #7917 )
2018-05-29 14:38:56 -07:00
find_duplicate_elements_op.h
fix -Wsign-compare warnings for some files inside c2 ( #18123 )
2019-03-19 10:39:20 -07:00
find_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
find_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
find_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
flatten_op.cc
add tensor and cost inference functions ( #17684 )
2019-03-06 23:34:02 -08:00
flatten_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
flexible_top_k.cc
Tensor construction: combine Resize+mutable_data - 2/4 ( #14205 )
2018-11-30 10:46:58 -08:00
flexible_top_k.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
floor_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
floor_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
floor_op.h
Tensor construction codemod(ResizeLike) - 4/7 ( #15088 )
2018-12-13 13:39:56 -08:00
free_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
free_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
free_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
fully_connected_op_gpu.cc
Integrate rocBLAS fp16 api into Caffe2 ( #14882 )
2018-12-10 17:54:06 -08:00
fully_connected_op.cc
Typo fix ( #18089 )
2019-03-16 15:07:01 -07:00
fully_connected_op.h
Int8 FC performance debugging ( #17700 )
2019-03-08 19:03:54 -08:00
fused_rowwise_8bit_conversion_ops.cc
move half<->float conversions to oss operators ( #17548 )
2019-03-07 13:00:13 -08:00
fused_rowwise_8bit_conversion_ops.h
change the epsilon for fp32/fp16 to uint8 to be the same ( #17062 )
2019-02-15 18:33:37 -08:00
fused_rowwise_random_quantization_ops.cc
minimize header file includes from _avx2.cc ( #14950 )
2018-12-13 00:18:11 -08:00
fused_rowwise_random_quantization_ops.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
gather_fused_8bit_rowwise_op.cc
gather_fused_8bit_rowwise_op.h
Tensor construction: combine Resize+mutable_data - 2/4 ( #14205 )
2018-11-30 10:46:58 -08:00
gather_op.cc
Implements Gather operator for arbitrary axis, sharing the code with BatchGather. ( #13756 )
2018-12-04 11:54:28 -08:00
gather_op.cu
Implements Gather operator for arbitrary axis, sharing the code with BatchGather. ( #13756 )
2018-12-04 11:54:28 -08:00
gather_op.cuh
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
gather_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
gather_ranges_to_dense_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
gather_ranges_to_dense_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
gelu_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
gelu_op.cu
use C10_REGISTER for GELU op
2019-04-12 11:41:04 -07:00
gelu_op.h
use C10_REGISTER for GELU op
2019-04-12 11:41:04 -07:00
generate_proposals_op_gpu_test.cc
Revert D14446895: [C2] Implement rotated generate_proposals_op without opencv dependency (~2x faster)
2019-03-23 09:38:55 -07:00
generate_proposals_op_test.cc
Implement rotated generate_proposals_op without opencv dependency (CPU version)
2019-03-28 17:02:50 -07:00
generate_proposals_op_util_boxes_test.cc
Removed obsolete argument correct_transform_coords in bbox_transform op. ( #16723 )
2019-02-20 13:22:33 -08:00
generate_proposals_op_util_boxes.h
Removed obsolete argument correct_transform_coords in bbox_transform op. ( #16723 )
2019-02-20 13:22:33 -08:00
generate_proposals_op_util_nms_gpu_test.cc
Implement rotated generate_proposals_op without opencv dependency (CPU version)
2019-03-28 17:02:50 -07:00
generate_proposals_op_util_nms_gpu.cu
Rotated boxes support for GPU GenerateProposals op ( #15470 )
2018-12-26 18:03:56 -08:00
generate_proposals_op_util_nms_gpu.h
Rotated boxes support for GPU GenerateProposals op ( #15470 )
2018-12-26 18:03:56 -08:00
generate_proposals_op_util_nms_test.cc
Implement rotated generate_proposals_op without opencv dependency (CPU version)
2019-03-28 17:02:50 -07:00
generate_proposals_op_util_nms.h
Implement rotated generate_proposals_op without opencv dependency (CPU version)
2019-03-28 17:02:50 -07:00
generate_proposals_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
generate_proposals_op.cu
Expose cuda kernel for caffe2::GenerateProposals
2019-03-04 14:59:08 -08:00
generate_proposals_op.h
Expose cuda kernel for caffe2::GenerateProposals
2019-03-04 14:59:08 -08:00
given_tensor_byte_string_to_uint8_fill_op.cc
Export uint8 tensors as byte string in mobile_exporter and add GivenTensorByteStringToUInt8FillOp ( #10385 )
2018-08-15 14:26:50 -07:00
given_tensor_byte_string_to_uint8_fill_op.cu
Export uint8 tensors as byte string in mobile_exporter and add GivenTensorByteStringToUInt8FillOp ( #10385 )
2018-08-15 14:26:50 -07:00
given_tensor_byte_string_to_uint8_fill_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
given_tensor_fill_op.cc
Updates to caffe2 operator documentation ( #7917 )
2018-05-29 14:38:56 -07:00
given_tensor_fill_op.cu
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
given_tensor_fill_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
glu_op.cc
glu_op.cu
glu_op.h
refactor caffe2 operator constructors - 3/9 ( #17084 )
2019-02-28 14:13:17 -08:00
group_norm_op.cc
correct comments in group_norm_op ( #19621 )
2019-04-23 13:31:15 -07:00
group_norm_op.cu
correct comments in group_norm_op ( #19621 )
2019-04-23 13:31:15 -07:00
group_norm_op.h
Optimize group_norm_op ( #17945 )
2019-03-21 13:05:01 -07:00
gru_unit_op_gpu.cu
HIP Operators Generator--> HipOpG ( #9322 )
2018-07-19 00:26:06 -07:00
gru_unit_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
gru_unit_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
h_softmax_op.cc
Tensor reinitialization codemod - 3/5 ( #15912 )
2019-01-16 19:49:01 -08:00
h_softmax_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
half_float_ops_test.cc
move half<->float conversions to oss operators ( #17548 )
2019-03-07 13:00:13 -08:00
half_float_ops.cc
Fix half-float conversion ops to handle tensors larger than 2B of params ( #17952 )
2019-03-12 23:03:22 -07:00
half_float_ops.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
half_float_ops.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
hard_sigmoid_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
hard_sigmoid_op.cu
Add GPU version of HardSigmoid Op to Caffe2 ( #10955 )
2018-08-29 14:55:29 -07:00
hard_sigmoid_op.h
Add CPU version of hard sigmoid operator to caffe2 ( #10837 )
2018-08-28 14:55:49 -07:00
heatmap_max_keypoint_op.cc
Revert D13866214: [c10] Expose HeatmapMaxKeypoints to torch
2019-02-06 08:26:16 -08:00
heatmap_max_keypoint_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
if_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
if_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
if_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
im2col_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
im2col_op.cc
bugfix im2col op
2018-03-30 21:00:44 -07:00
im2col_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
index_hash_ops.cc
Remove many caffe2::TIndex and replace them with int64_t ( #11943 )
2018-09-22 18:11:04 -07:00
index_hash_ops.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
index_ops.cc
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
index_ops.h
move structs to header file ( #14728 )
2018-12-04 00:42:41 -08:00
inference_lstm_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
inference_lstm_op.h
Delete C10Tensor ( #19328 )
2019-04-17 00:02:27 -07:00
instance_norm_gradient_op.cc
Tensor reinitialization codemod - 3/5 ( #15912 )
2019-01-16 19:49:01 -08:00
instance_norm_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
instance_norm_op.cu
Tensor method rename ndim()->dim() - 2/3 ( #16679 )
2019-02-04 11:12:57 -08:00
instance_norm_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
integral_image_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
integral_image_op.cu
Tensor method rename ndim()->dim() - 2/3 ( #16679 )
2019-02-04 11:12:57 -08:00
integral_image_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
is_empty_op.cc
Remove BUILD_CAFFE2 and build everything ( #8338 )
2018-08-31 13:10:24 -07:00
is_empty_op.h
Tensor construction: combine Resize+mutable_data - 2/4 ( #14205 )
2018-11-30 10:46:58 -08:00
jsd_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
jsd_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
key_split_ops.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
key_split_ops.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
last_n_window_collector.cc
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
layer_norm_op.cc
Add elementwise_affine for layer_norm_op ( #19713 )
2019-04-26 17:20:01 -07:00
layer_norm_op.cu
Add elementwise_affine for layer_norm_op ( #19713 )
2019-04-26 17:20:01 -07:00
layer_norm_op.h
Add elementwise_affine for layer_norm_op ( #19713 )
2019-04-26 17:20:01 -07:00
leaky_relu_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
leaky_relu_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
leaky_relu_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
length_split_op.cc
Add new LengthsSplit operator ( #10974 )
2018-09-10 15:40:28 -07:00
length_split_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
lengths_pad_op.cc
Update from facebook ( #6871 )
2018-04-23 15:01:56 -07:00
lengths_pad_op.cu
Update from facebook ( #6871 )
2018-04-23 15:01:56 -07:00
lengths_pad_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
lengths_reducer_fused_8bit_rowwise_ops.cc
Add filler for SparseLengthsWeightedSum ( #13949 )
2018-11-16 11:31:05 -08:00
lengths_reducer_fused_8bit_rowwise_ops.h
Tensor construction: combine Resize+mutable_data - 2/4 ( #14205 )
2018-11-30 10:46:58 -08:00
lengths_reducer_ops.cc
Add onnxifi support to SparseLengthsWeightedSum ( #14210 )
2018-11-21 15:47:24 -08:00
lengths_reducer_ops.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
lengths_reducer_rowwise_8bit_ops.cc
Move registry fully to c10 ( #12077 )
2018-09-27 03:09:54 -07:00
lengths_reducer_rowwise_8bit_ops.h
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
lengths_tile_op.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
lengths_tile_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
lengths_tile_op.h
Tensor reinitialization codemod - 3/5 ( #15912 )
2019-01-16 19:49:01 -08:00
lengths_top_k_op.cc
Tensor construction: combine Resize+mutable_data - 2/4 ( #14205 )
2018-11-30 10:46:58 -08:00
lengths_top_k_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
listwise_l2r_op.cc
Tensor reinitialization codemod - 3/5 ( #15912 )
2019-01-16 19:49:01 -08:00
listwise_l2r_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
load_save_op_gpu.cc
Rename cuda_gpu_id to device_id in DeviceOption ( #12456 )
2018-10-09 15:54:04 -07:00
load_save_op.cc
Make chunk size configurable in SaveOp ( #12949 )
2018-10-25 15:47:34 -07:00
load_save_op.h
Avoid Output Uninitialized Blobs in Load with load_all=1 ( #19133 )
2019-04-27 10:45:44 -07:00
local_response_normalization_op_cudnn.cc
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
local_response_normalization_op.cc
Move math::Axpy function to elementwise lib ( #18316 )
2019-03-26 12:19:19 -07:00
local_response_normalization_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
local_response_normalization_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
locally_connected_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
locally_connected_op_impl.h
Back out "[pt1][tensor] Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize" ( #16516 )
2019-01-30 12:50:38 -08:00
locally_connected_op_util.cc
Update Im2ColNd functions ( #7505 )
2018-05-12 15:59:50 -07:00
locally_connected_op_util.h
Additional changes to make GPU builds work ( #10507 )
2018-08-16 13:25:27 -07:00
locally_connected_op.cc
Move exception to C10 ( #12354 )
2018-10-15 13:33:18 -07:00
locally_connected_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
log_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
log_op.cc
Revert D10439558: Add cost for non-linear ops
2018-11-16 23:30:05 -08:00
log_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
logit_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
logit_op.cu
Remove template parameter from Tensor ( #9939 )
2018-07-27 10:56:39 -07:00
logit_op.h
refactor caffe2 operator constructors - 4/9 ( #17085 )
2019-02-28 14:23:52 -08:00
loss_op.cc
Updates to caffe2 operator documentation ( #7917 )
2018-05-29 14:38:56 -07:00
loss_op.cu
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
loss_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
lp_pool_op.cc
Back out "[pt1][tensor] Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize" ( #16516 )
2019-01-30 12:50:38 -08:00
lp_pool_op.cu
Tensor method rename size()->numel() - 1/3
2019-02-04 23:33:17 -08:00
lpnorm_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
lpnorm_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
lstm_unit_op_gpu.cu
codemod: caffe::float16 -> at::Half ( #11785 )
2018-09-20 18:55:19 -07:00
lstm_unit_op.cc
Revert D10439558: Add cost for non-linear ops
2018-11-16 23:30:05 -08:00
lstm_unit_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
lstm_utils.h
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
map_ops.cc
Add interface to provide blob types to shape&type inference ( #9643 )
2018-07-24 11:58:05 -07:00
map_ops.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
margin_ranking_criterion_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
margin_ranking_criterion_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
margin_ranking_criterion_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
matmul_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
matmul_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
matmul_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
max_pool_with_index_gpu.h
HIP Operators Generator--> HipOpG ( #9322 )
2018-07-19 00:26:06 -07:00
max_pool_with_index.cu
Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize ( #17764 )
2019-03-07 18:38:53 -08:00
mean_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
mean_op.cu
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
mean_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
mem_query_op.cu
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
merge_id_lists_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
merge_id_lists_op.h
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
minmax_gradient_ops.cc
Separate elementwise level2 math functions ( #16753 )
2019-02-07 18:38:26 -08:00
minmax_ops.cc
Separate elementwise level2 math functions ( #16753 )
2019-02-07 18:38:26 -08:00
minmax_ops.cu
Separate elementwise level2 math functions ( #16753 )
2019-02-07 18:38:26 -08:00
minmax_ops.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
mod_op.cc
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
mod_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
moments_op.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
moments_op.cu
Use FixedDivisor in Reduce and Broadcast CUDA kernels ( #9072 )
2018-07-01 00:25:34 -07:00
moments_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
multi_class_accuracy_op.cc
Tensor construction: combine Resize+mutable_data - 3/4 ( #13944 )
2018-11-19 15:28:13 -08:00
multi_class_accuracy_op.cu
Tensor method rename ndim()->dim() - 2/3 ( #16679 )
2019-02-04 11:12:57 -08:00
multi_class_accuracy_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
negate_gradient_op_gpu.cc
negate_gradient_op.cc
negate_gradient_op.h
Shut up "address will always evaluate to 'true'" warnings ( #14774 )
2018-12-05 21:18:31 -08:00
negative_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
negative_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
negative_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
ngram_ops.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
ngram_ops.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
no_default_engine_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
norm_planar_yuv_op.cc
Apply modernize-use-override - 2/2
2019-02-13 21:01:28 -08:00
normalize_l1_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
normalize_l1_op.h
Tensor construction codemod(ResizeLike) - 5/7 ( #15084 )
2018-12-13 12:42:52 -08:00
normalize_op.cc
Fix compilation error ( #17860 )
2019-03-11 10:26:42 -07:00
normalize_op.h
Fix compilation error ( #17860 )
2019-03-11 10:26:42 -07:00
normalize_ops.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
numpy_tile_op.cc
[caffe2][ONNX] Implement CPU NumpyTileOp and corresponding ONNX backend ( #7053 )
2018-04-27 19:58:15 -07:00
numpy_tile_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
one_hot_ops.cc
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
one_hot_ops.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
one_hot_ops.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
onnx_while_op.cc
[caffe2] Add flag to ONNXWhile to skip scoping ( #6910 )
2018-04-24 16:53:22 -07:00
onnx_while_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
onnxifi_op.cc
Automatic update of fbcode/foxi to 8f74bc4df3a4cfc69b1a3eadf62aa29d9961c72d AND update Glow AND update C2 ( #19792 )
2019-04-25 21:03:32 -07:00
onnxifi_op.h
Surface the Glow traces to C2 ( #19087 )
2019-04-23 12:27:49 -07:00
op_utils_cudnn.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
operator_fallback_gpu_test.cc
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
operator_fallback_gpu.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
order_switch_ops_cudnn.cc
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
order_switch_ops_gpu.cc
Optimize NCHW2NHWC on GPU ( #12910 )
2018-10-22 11:24:29 -07:00
order_switch_ops.cc
Optimize NCHW2NHWC on GPU ( #12910 )
2018-10-22 11:24:29 -07:00
order_switch_ops.h
Tensor construction: combine Resize+mutable_data - 3/4 ( #13944 )
2018-11-19 15:28:13 -08:00
pack_rnn_sequence_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
pack_rnn_sequence_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
pack_segments.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
pack_segments.cu
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
pack_segments.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
pad_op_gpu.cu
Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize ( #17764 )
2019-03-07 18:38:53 -08:00
pad_op.cc
Back out "[pt1][tensor] Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize" ( #16516 )
2019-01-30 12:50:38 -08:00
pad_op.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
partition_ops.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
partition_ops.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
percentile_op.cc
Tensor reinitialization codemod - 3/5 ( #15912 )
2019-01-16 19:49:01 -08:00
percentile_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
perplexity_op.cc
Tensor construction: combine Resize+mutable_data - 3/4 ( #13944 )
2018-11-19 15:28:13 -08:00
perplexity_op.cu
Tensor construction codemod ( #16568 )
2019-02-05 18:51:02 -08:00
perplexity_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
piecewise_linear_transform_op.cc
Remove template parameter from Tensor ( #9939 )
2018-07-27 10:56:39 -07:00
piecewise_linear_transform_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
piecewise_linear_transform_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
pool_gradient_op.cc
Add count_include_pad to average_pool_gradient_op ( #15997 )
2019-01-15 16:56:40 -08:00
pool_op_cudnn.cc
Change ConvPoolOp<Context>::SetOutputSize to ConvPoolOp<Context>::GetOutputSize ( #17764 )
2019-03-07 18:38:53 -08:00
pool_op_util.cc
Add count_include_pad arg for PoolOpGradient on CPU and fix ARM performance issue. ( #15651 )
2019-01-03 00:18:47 -08:00
pool_op_util.h
Add count_include_pad arg for PoolOpGradient on CPU and fix ARM performance issue. ( #15651 )
2019-01-03 00:18:47 -08:00
pool_op.cc
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
pool_op.cu
Separate reduce functions from math ( #16929 )
2019-02-13 17:50:47 -08:00
pool_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
pow_op.cc
Revert D10439558: Add cost for non-linear ops
2018-11-16 23:30:05 -08:00
pow_op.cu
Hipify caffe2/utils/math_gpu.cu ( #13521 )
2018-11-07 11:34:15 -08:00
pow_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
prefetch_op.h
Replace SwitchToDevice(0) with SwitchToDevice() ( #15126 )
2018-12-17 15:15:00 -08:00
prelu_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
prelu_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
prelu_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
prepend_dim_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
prepend_dim_op.cc
Support conversion from Caffe2 MergeDim to ONNX Reshape + Squeeze. ( #16189 )
2019-02-13 15:53:38 -08:00
prepend_dim_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
quant_decode_op.cc
Fix include paths for typeid.h ( #13689 )
2018-11-14 18:04:09 -08:00
quant_decode_op.h
refactor caffe2 operator constructors - 6/9 ( #17087 )
2019-02-28 14:23:57 -08:00
rank_loss_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
rank_loss_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
reciprocal_gradient_op.cc
Adding reciprocal operator and a test
2018-07-27 18:24:43 -07:00
reciprocal_op.cc
Adding reciprocal operator and a test
2018-07-27 18:24:43 -07:00
reciprocal_op.cu
Adding reciprocal operator and a test
2018-07-27 18:24:43 -07:00
reciprocal_op.h
Adding reciprocal operator and a test
2018-07-27 18:24:43 -07:00
reduce_front_back_max_ops.cc
Split reduction_front_backops.[cc|cu] into smaller units to allow build of smaller size ( #12315 )
2018-10-05 16:50:21 -07:00
reduce_front_back_max_ops.cu
Split reduction_front_backops.[cc|cu] into smaller units to allow build of smaller size ( #12315 )
2018-10-05 16:50:21 -07:00
reduce_front_back_max_ops.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
reduce_front_back_mean_ops.cc
Export ReduceMean/ReduceFrontMean/ReduceBackMean (Caffe2) to ReduceMean (ONNX). ( #16727 )
2019-02-12 13:35:32 -08:00
reduce_front_back_sum_mean_ops.cu
Split reduction_front_backops.[cc|cu] into smaller units to allow build of smaller size ( #12315 )
2018-10-05 16:50:21 -07:00
reduce_front_back_sum_mean_ops.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
reduce_front_back_sum_ops.cc
Split reduction_front_backops.[cc|cu] into smaller units to allow build of smaller size ( #12315 )
2018-10-05 16:50:21 -07:00
reduce_ops.cc
Export ReduceMean/ReduceFrontMean/ReduceBackMean (Caffe2) to ReduceMean (ONNX). ( #16727 )
2019-02-12 13:35:32 -08:00
reduce_ops.cu
Use FixedDivisor in Reduce and Broadcast CUDA kernels ( #9072 )
2018-07-01 00:25:34 -07:00
reduce_ops.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
reducer_functors.h
Rename IntList to IntArrayRef. ( #16751 )
2019-02-05 14:54:34 -08:00
reduction_ops.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
reduction_ops.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
reduction_ops.h
Reapply D14078519 ( #17596 )
2019-03-06 13:51:00 -08:00
relu_n_op.cc
Remove template parameter from Tensor ( #9939 )
2018-07-27 10:56:39 -07:00
relu_n_op.cu
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
relu_n_op.h
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
relu_op.cc
Use REGISTER_CPU_GRADIENT_OPERATOR for many operators ( #12616 )
2018-10-24 13:48:17 -07:00
relu_op.cu
Fix relu bug for empty tensor ( #19451 )
2019-04-19 15:21:07 -07:00
relu_op.h
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
remove_data_blocks_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
remove_data_blocks_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
replace_nan_op.cc
Remove many caffe2::TIndex and replace them with int64_t ( #11943 )
2018-09-22 18:11:04 -07:00
replace_nan_op.cu
Remove many caffe2::TIndex and replace them with int64_t ( #11943 )
2018-09-22 18:11:04 -07:00
replace_nan_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
reservoir_sampling.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
reshape_op_gpu_test.cc
Rename ndim() -> dim() - 5/6
2018-11-06 16:38:35 -08:00
reshape_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
reshape_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
reshape_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
resize_op.cc
Add NHWC support to Resize Operator ( #15553 )
2019-01-08 16:44:17 -08:00
resize_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
resize_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
reverse_packed_segs_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
reverse_packed_segs_op.cu
Tensor method rename ndim()->dim() - 2/3 ( #16679 )
2019-02-04 11:12:57 -08:00
reverse_packed_segs_op.h
Tensor construction: combine Resize+mutable_data - 3/4 ( #13944 )
2018-11-19 15:28:13 -08:00
rmac_regions_op.cc
Remove Context dependency from Tensor class ( #14269 )
2018-11-28 15:45:38 -08:00
rmac_regions_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
rmac_regions_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
roi_align_gradient_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
roi_align_gradient_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
roi_align_gradient_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
roi_align_op_gpu_test.cc
Remove calls to CopyFrom that can be sync ( #13205 )
2018-10-29 13:57:42 -07:00
roi_align_op.cc
Use string based schema for exposing caffe2 ops ( #19287 )
2019-04-18 02:04:50 -07:00
roi_align_op.cu
register RoIAlign with C10
2019-03-14 11:55:29 -07:00
roi_align_op.h
register RoIAlign with C10
2019-03-14 11:55:29 -07:00
roi_align_rotated_gradient_op.cc
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
roi_align_rotated_gradient_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
roi_align_rotated_gradient_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
roi_align_rotated_op.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
roi_align_rotated_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
roi_align_rotated_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
roi_pool_op.cc
Tensor dims() -> sizes() (caffe2/operators) - 4/5 ( #13031 )
2018-10-24 15:07:42 -07:00
roi_pool_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
roi_pool_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
rowmul_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
rowmul_op.h
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
rsqrt_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
rsqrt_op.cu
Separate elementwise level2 math functions ( #16753 )
2019-02-07 18:38:26 -08:00
rsqrt_op.h
Add Cube and Cbrt Ops ( #8991 )
2018-06-28 14:55:30 -07:00
scale_op_gpu.cc
codemod: caffe::float16 -> at::Half ( #11785 )
2018-09-20 18:55:19 -07:00
scale_op.cc
Remove template parameter from Tensor ( #9939 )
2018-07-27 10:56:39 -07:00
scale_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
segment_reduction_op_gpu.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
segment_reduction_op.cc
Windows shared build ( #13550 )
2018-11-16 12:16:28 -08:00
segment_reduction_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
selu_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
selu_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
selu_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
sequence_ops.cc
Undefined behavior with memset of std::string to 0 ( #18703 )
2019-04-02 10:10:11 -07:00
sequence_ops.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
sequence_ops.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
shape_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
shape_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
shape_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
sigmoid_gradient_op.cc
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
sigmoid_op_cudnn.cc
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
sigmoid_op.cc
Revert D10439558: Add cost for non-linear ops
2018-11-16 23:30:05 -08:00
sigmoid_op.cu
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
sigmoid_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
sin_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
sin_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
sin_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
sinh_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
sinh_op.cu
Implement Sinh and Cosh ( #9213 )
2018-07-10 18:55:31 -07:00
sinh_op.h
Implement Sinh and Cosh ( #9213 )
2018-07-10 18:55:31 -07:00
sinusoid_position_encoding_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
sinusoid_position_encoding_op.h
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
slice_op.cc
Use REGISTER_CPU_GRADIENT_OPERATOR for many operators ( #12616 )
2018-10-24 13:48:17 -07:00
slice_op.cu
refactor caffe2 operator constructors - 7/9 ( #17088 )
2019-02-28 14:23:53 -08:00
slice_op.h
fix -Wsign-compare warnings for some files inside c2 ( #18123 )
2019-03-19 10:39:20 -07:00
softmax_op_cudnn.cc
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
softmax_op.cc
Optimize SoftmaxOp on CPU ( #18635 )
2019-04-10 18:52:15 -07:00
softmax_op.h
Optimize SoftmaxOp on CPU ( #18635 )
2019-04-10 18:52:15 -07:00
softmax_ops.cu
Allow extracting element-wise loss in softmax ( #19579 )
2019-04-23 11:49:49 -07:00
softmax_utils.cc
Optimize SoftmaxOp on CPU ( #18635 )
2019-04-10 18:52:15 -07:00
softmax_utils.h
Optimize SoftmaxOp on CPU ( #18635 )
2019-04-10 18:52:15 -07:00
softmax_with_loss_op.cc
Allow extracting element-wise loss in softmax ( #19579 )
2019-04-23 11:49:49 -07:00
softmax_with_loss_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
softplus_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
softplus_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
softplus_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
softsign_op.cc
Use REGISTER_CPU_GRADIENT_OPERATOR for many operators ( #12616 )
2018-10-24 13:48:17 -07:00
softsign_op.cu
Add math functions overloaded over different numeric types for cuda and hip ( #13602 )
2018-11-06 01:40:31 -08:00
softsign_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
space_batch_op_gpu.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
space_batch_op.cc
Updates to caffe2 operator documentation ( #7917 )
2018-05-29 14:38:56 -07:00
space_batch_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
sparse_normalize_op_gpu.cu
Windows DLL build with Caffe2 code ( #11266 )
2018-09-06 15:12:20 -07:00
sparse_normalize_op.cc
Rename ndim() -> dim() - 5/6
2018-11-06 16:38:35 -08:00
sparse_normalize_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
sparse_to_dense_mask_op.cc
Windows raw string fix ( #10998 )
2018-08-29 11:40:08 -07:00
sparse_to_dense_mask_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
sparse_to_dense_op.cc
Windows raw string fix ( #10998 )
2018-08-29 11:40:08 -07:00
sparse_to_dense_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
sparse_to_dense_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
spatial_batch_norm_gradient_op.cc
Optimize SpatialBNOp on GPU ( #16395 )
2019-01-28 09:36:45 -08:00
spatial_batch_norm_op_cudnn.cu
Optimize SpatialBNOp on GPU ( #16395 )
2019-01-28 09:36:45 -08:00
spatial_batch_norm_op_impl.cuh
Separate gpu reduce functions ( #17146 )
2019-02-20 14:49:01 -08:00
spatial_batch_norm_op.cc
Revert D13747581: Optimize SpatialBN on GPU
2019-01-24 15:26:37 -08:00
spatial_batch_norm_op.cu
Optimize SpatialBNOp on GPU ( #16395 )
2019-01-28 09:36:45 -08:00
spatial_batch_norm_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
spatial_softmax_with_loss_op.cc
Optimize SoftmaxOp on CPU ( #18635 )
2019-04-10 18:52:15 -07:00
spatial_softmax_with_loss_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
sqr_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
sqr_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
sqr_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
sqrt_op_gpu.cc
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
sqrt_op.cc
Update operator documentation with markdown descriptions and interfaces ( #8085 )
2018-06-15 19:02:24 -04:00
sqrt_op.h
Seperate level1 elementwise functions from math ( #16397 )
2019-01-30 00:04:12 -08:00
square_root_divide_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
square_root_divide_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
stats_ops.cc
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
stats_put_ops.cc
Make Stat put ops accept empty tensors safely ( #13178 )
2018-10-30 13:28:58 -07:00
stats_put_ops.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
stop_gradient_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
stop_gradient.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
stop_gradient.h
Remove Context dependency from Tensor class ( #14269 )
2018-11-28 15:45:38 -08:00
string_ops_test.cc
Remove caffe2::ShareData ( #16139 )
2019-01-25 15:39:11 -08:00
string_ops.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
string_ops.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
stump_func_op.cc
Tensor construction codemod(ResizeLike) - 6/7 ( #15137 )
2018-12-13 12:47:33 -08:00
stump_func_op.cu
Tensor method rename size()->numel() - 2/3 ( #16745 )
2019-02-04 23:59:10 -08:00
stump_func_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
stylizer_ops.cc
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
summarize_op.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
summarize_op.cu
Tensor method rename size()->numel() - 3/3 ( #16747 )
2019-02-04 23:54:33 -08:00
summarize_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
swish_op.cc
Tensor construction codemod(ResizeLike) - 7/7 ( #15087 )
2018-12-20 15:33:07 -08:00
swish_op.cu
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
swish_op.h
[Caffe2] Update elementwise ops to support numpy style boradcast ( #8070 )
2018-06-05 15:49:16 -07:00
tan_op.cc
Fix Eigen issue on OS X with CUDA and nvcc compile ( #9350 )
2018-07-11 14:00:05 -07:00
tan_op.cu
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
tan_op.h
Update from Facebook ( #8887 )
2018-06-26 14:55:48 -07:00
tanh_gradient_op.cc
Fix TanhGradientOperator linker errors ( #10426 )
2018-08-13 17:57:10 -07:00
tanh_op_cudnn.cc
Add cudnn activation ops ( #9379 )
2018-07-12 23:18:56 -07:00
tanh_op.cc
Revert D10439558: Add cost for non-linear ops
2018-11-16 23:30:05 -08:00
tanh_op.cu
Move tanh function to math ( #9328 )
2018-07-11 13:59:50 -07:00
tanh_op.h
Move tanh function to math ( #9328 )
2018-07-11 13:59:50 -07:00
tensor_protos_db_input_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
tensor_protos_db_input.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
tensor_protos_db_input.h
Remove partially initialized Tensor in Deserialization ( #14197 )
2018-12-10 17:17:29 -08:00
text_file_reader_utils_test.cc
Apply modernize-use-override - 2/2
2019-02-13 21:01:28 -08:00
text_file_reader_utils.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
text_file_reader_utils.h
Remove BUILD_CAFFE2 and build everything ( #8338 )
2018-08-31 13:10:24 -07:00
text_file_reader.cc
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
thresholded_relu_op.cc
Tensor construction codemod(ResizeLike) - 7/7 ( #15087 )
2018-12-20 15:33:07 -08:00
thresholded_relu_op.cu
Tensor method rename size()->numel() - 3/3 ( #16747 )
2019-02-04 23:54:33 -08:00
thresholded_relu_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
tile_op.cc
Temp fix for TileOp backward compatibility ( #18026 )
2019-03-14 13:54:29 -07:00
tile_op.cu
Optimize TileOp ( #17290 )
2019-03-12 12:16:30 -07:00
tile_op.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
top_k_heap_selection.cuh
Do not apply these explicit unroll pragmas for ROCm. ( #18204 )
2019-03-20 08:06:07 -07:00
top_k_radix_selection.cuh
Separate affine_channel from math and optimize it ( #16135 )
2019-01-18 22:40:16 -08:00
top_k.cc
Rename IntList to IntArrayRef. ( #16751 )
2019-02-05 14:54:34 -08:00
top_k.cu
Rename IntList to IntArrayRef. ( #16751 )
2019-02-05 14:54:34 -08:00
top_k.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
transpose_op_cudnn.cc
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
transpose_op.cc
Simplify InheritOnnxSchema registration ( #12696 )
2018-10-16 19:59:49 -07:00
transpose_op.cu
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
transpose_op.h
Update math::Transpose to support tensor with size > 2G ( #17670 )
2019-03-20 18:22:21 -07:00
tt_linear_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
tt_linear_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
unique_ops.cc
Tensor construction codemod(ResizeLike) - 7/7 ( #15087 )
2018-12-20 15:33:07 -08:00
unique_ops.cu
Tensor construction codemod ( #16568 )
2019-02-05 18:51:02 -08:00
unique_ops.h
Tensor reinitialization codemod - 4/5 ( #15967 )
2019-01-11 16:41:19 -08:00
upsample_op.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
upsample_op.cu
Tensor method rename size()->numel() - 3/3 ( #16747 )
2019-02-04 23:54:33 -08:00
upsample_op.h
refactor caffe2 operator constructors - 8/9 ( #17089 )
2019-02-28 14:45:20 -08:00
utility_ops_gpu_test.cc
Rename ndim() -> dim() - 5/6
2018-11-06 16:38:35 -08:00
utility_ops_test.cc
Rename ndim() -> dim() - 5/6
2018-11-06 16:38:35 -08:00
utility_ops.cc
ONNX Export IsNan op
2019-03-15 12:19:03 -07:00
utility_ops.cu
Separate elementwise level2 math functions ( #16753 )
2019-02-07 18:38:26 -08:00
utility_ops.h
Revert "Tensor construction codemod(raw_mutable_data) ( #16373 )" ( #18680 )
2019-04-01 14:39:13 -07:00
variable_length_sequence_padding.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
variable_length_sequence_padding.h
refactor caffe2 operator constructors - 9/9 ( #17090 )
2019-02-28 09:53:18 -08:00
weighted_multi_sampling_op.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
weighted_multi_sampling_op.h
refactor caffe2 operator constructors - 9/9 ( #17090 )
2019-02-28 09:53:18 -08:00
weighted_sample_op.cc
Tensor construction: combine Resize+mutable_data - 4/4 ( #13856 )
2018-11-27 12:34:25 -08:00
weighted_sample_op.cu
Tensor reinitialization codemod - 4/5 ( #15967 )
2019-01-11 16:41:19 -08:00
weighted_sample_op.h
refactor caffe2 operator constructors - 9/9 ( #17090 )
2019-02-28 09:53:18 -08:00
while_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
while_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
while_op.h
refactor caffe2 operator constructors - 9/9 ( #17090 )
2019-02-28 09:53:18 -08:00
workspace_ops.cc
refactor caffe2 operator constructors - 9/9 ( #17090 )
2019-02-28 09:53:18 -08:00
zero_gradient_op_gpu.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
zero_gradient_op.cc
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00
zero_gradient_op.h
Remove Apache headers from source.
2018-03-27 13:10:18 -07:00