pytorch/caffe2/operators/normalize_l1_op.h
Vitaly Fedyunin a1c234e372 Revert D22330340: [C2] Fixed a bug in normalization operator
Test Plan: revert-hammer

Differential Revision:
D22330340 (ce63f70981)

Original commit changeset: 0bccf925bb76

fbshipit-source-id: e27d70dee0fbe9e708b0cf3be81dbd33c4015026
2020-07-02 16:05:23 -07:00

40 lines
1.0 KiB
C++

#ifndef CAFFE2_OPERATORS_NORMALIZE_L1_OP_H_
#define CAFFE2_OPERATORS_NORMALIZE_L1_OP_H_
#include "caffe2/core/context.h"
#include "caffe2/core/operator.h"
#include "caffe2/utils/math.h"
namespace caffe2 {
template <typename T, class Context>
class NormalizeL1Op final : public Operator<Context> {
public:
USE_OPERATOR_CONTEXT_FUNCTIONS;
USE_SIMPLE_CTOR_DTOR(NormalizeL1Op)
bool RunOnDevice() override {
const auto& x = Input(0);
const auto* xData = x.template data<T>();
auto* y = Output(0, x.sizes(), at::dtype<T>());
auto* yData = y->template mutable_data<T>();
const auto canonical_axis = x.canonical_axis_index(
this->template GetSingleArgument<int>("axis", -1));
const int m = x.dim32(canonical_axis);
const int n = x.numel() / m;
const int sf = x.size_from_dim(canonical_axis + 1);
DoNormalize(xData, yData, m, n, sf);
return true;
}
private:
void
DoNormalize(const T* xData, T* yData, const int m, const int n, const int sf);
};
} // namespace caffe2
#endif // CAFFE2_OPERATORS_NORMALIZE_L1_OP_H_