Automated Code Change

PiperOrigin-RevId: 825996910
This commit is contained in:
A. Unique TensorFlower 2025-10-30 06:01:51 -07:00 committed by TensorFlower Gardener
parent 3449303622
commit 0a1309a2e5
8 changed files with 53 additions and 46 deletions

View File

@ -82,8 +82,8 @@ class SqlDatasetOp : public DatasetOpKernel {
private:
class Dataset : public DatasetBase {
public:
Dataset(OpKernelContext* ctx, const string& driver_name,
const string& data_source_name, const string& query,
Dataset(OpKernelContext* ctx, const std::string& driver_name,
const std::string& data_source_name, const std::string& query,
const DataTypeVector& output_types,
const std::vector<PartialTensorShape>& output_shapes)
: DatasetBase(DatasetContext(ctx)),
@ -94,7 +94,7 @@ class SqlDatasetOp : public DatasetOpKernel {
output_shapes_(output_shapes) {}
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::Sql")});
}
@ -107,7 +107,7 @@ class SqlDatasetOp : public DatasetOpKernel {
return output_shapes_;
}
string DebugString() const override { return "SqlDatasetOp::Dataset"; }
std::string DebugString() const override { return "SqlDatasetOp::Dataset"; }
absl::Status InputDatasets(
std::vector<const DatasetBase*>* inputs) const override {

View File

@ -44,9 +44,11 @@ static mutex* get_counters_map_lock() {
return &counters_map_lock;
}
static std::unordered_map<string, monitoring::Counter<1>*>* get_counters_map() {
static std::unordered_map<string, monitoring::Counter<1>*>* counters_map =
new std::unordered_map<string, monitoring::Counter<1>*>;
static std::unordered_map<std::string, monitoring::Counter<1>*>*
get_counters_map() {
static std::unordered_map<std::string, monitoring::Counter<1>*>*
counters_map =
new std::unordered_map<std::string, monitoring::Counter<1>*>;
return counters_map;
}
@ -54,7 +56,7 @@ class StatsAggregatorImpl : public StatsAggregator {
public:
StatsAggregatorImpl() {}
void AddToHistogram(const string& name, absl::Span<const double> values,
void AddToHistogram(const std::string& name, absl::Span<const double> values,
const int64_t steps) override {
mutex_lock l(mu_);
histogram::Histogram& histogram = histograms_[name];
@ -63,7 +65,7 @@ class StatsAggregatorImpl : public StatsAggregator {
}
}
void AddScalar(const string& name, float value,
void AddScalar(const std::string& name, float value,
const int64_t steps) override {
mutex_lock l(mu_);
scalars_[name] = value;
@ -72,7 +74,7 @@ class StatsAggregatorImpl : public StatsAggregator {
void EncodeToProto(Summary* out_summary) override {
mutex_lock l(mu_);
for (const auto& pair : histograms_) {
const string& name = pair.first;
const std::string& name = pair.first;
const histogram::Histogram& histogram = pair.second;
Summary::Value* value = out_summary->add_value();
@ -94,7 +96,7 @@ class StatsAggregatorImpl : public StatsAggregator {
return absl::OkStatus();
}
void IncrementCounter(const string& name, const string& label,
void IncrementCounter(const std::string& name, const std::string& label,
int64_t val) override {
mutex_lock l(*get_counters_map_lock());
auto counters_map = get_counters_map();
@ -112,9 +114,9 @@ class StatsAggregatorImpl : public StatsAggregator {
private:
mutex mu_;
std::unordered_map<string, histogram::Histogram> histograms_
std::unordered_map<std::string, histogram::Histogram> histograms_
TF_GUARDED_BY(mu_);
std::unordered_map<string, float> scalars_ TF_GUARDED_BY(mu_);
std::unordered_map<std::string, float> scalars_ TF_GUARDED_BY(mu_);
StatsAggregatorImpl(const StatsAggregatorImpl&) = delete;
void operator=(const StatsAggregatorImpl&) = delete;
};
@ -143,7 +145,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
}
}
void AddToHistogram(const string& name, absl::Span<const double> values,
void AddToHistogram(const std::string& name, absl::Span<const double> values,
const int64_t steps) override {
mutex_lock l(mu_);
histogram::Histogram& histogram = histograms_[name];
@ -153,7 +155,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
AddToEvents(name, steps, histogram);
}
void AddScalar(const string& name, float value,
void AddScalar(const std::string& name, float value,
const int64_t steps) override {
mutex_lock l(mu_);
AddToEvents(name, steps, value);
@ -167,7 +169,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
return absl::OkStatus();
}
void IncrementCounter(const string& name, const string& label,
void IncrementCounter(const std::string& name, const std::string& label,
int64_t val) override {
mutex_lock l(*get_counters_map_lock());
auto counters_map = get_counters_map();
@ -203,7 +205,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
}
private:
void AddToEvents(const string& name, const int64_t steps,
void AddToEvents(const std::string& name, const int64_t steps,
const float scalar_value) TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) {
if (summary_writer_interface_ == nullptr) {
return;
@ -218,7 +220,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
TF_CHECK_OK(summary_writer_interface_->WriteEvent(std::move(e)));
}
void AddToEvents(const string& name, const int64_t steps,
void AddToEvents(const std::string& name, const int64_t steps,
const histogram::Histogram& histogram)
TF_EXCLUSIVE_LOCKS_REQUIRED(mu_) {
if (summary_writer_interface_ == nullptr) {
@ -238,7 +240,7 @@ class StatsAggregatorImplV2 : public StatsAggregator {
nullptr;
// not owned, we might be associating the default summary_writer from the
// context
std::unordered_map<string, histogram::Histogram> histograms_
std::unordered_map<std::string, histogram::Histogram> histograms_
TF_GUARDED_BY(mu_);
StatsAggregatorImplV2(const StatsAggregatorImplV2&) = delete;
void operator=(const StatsAggregatorImplV2&) = delete;

View File

@ -58,7 +58,8 @@ class LatencyStatsDatasetOp : public UnaryDatasetOpKernel {
private:
class Dataset : public DatasetBase {
public:
explicit Dataset(OpKernelContext* ctx, const DatasetBase* input, string tag)
explicit Dataset(OpKernelContext* ctx, const DatasetBase* input,
std::string tag)
: DatasetBase(DatasetContext(ctx)),
input_(input),
tag_(std::move(tag)) {
@ -68,7 +69,7 @@ class LatencyStatsDatasetOp : public UnaryDatasetOpKernel {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::LatencyStats")});
}
@ -80,7 +81,7 @@ class LatencyStatsDatasetOp : public UnaryDatasetOpKernel {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "LatencyStatsDatasetOp::Dataset";
}
@ -125,10 +126,10 @@ class LatencyStatsDatasetOp : public UnaryDatasetOpKernel {
std::vector<Tensor>* out_tensors,
bool* end_of_sequence) override {
tf_shared_lock l(mu_);
uint64 start = EnvTime::NowMicros();
uint64_t start = EnvTime::NowMicros();
absl::Status s =
input_impl_->GetNext(ctx, out_tensors, end_of_sequence);
uint64 end = EnvTime::NowMicros();
uint64_t end = EnvTime::NowMicros();
auto stats_aggregator = ctx->stats_aggregator();
if (stats_aggregator && !*end_of_sequence) {
int64_t steps = num_elements();
@ -184,7 +185,8 @@ class BytesProducedStatsDatasetOp : public UnaryDatasetOpKernel {
private:
class Dataset : public DatasetBase {
public:
explicit Dataset(OpKernelContext* ctx, const DatasetBase* input, string tag)
explicit Dataset(OpKernelContext* ctx, const DatasetBase* input,
std::string tag)
: DatasetBase(DatasetContext(ctx)),
input_(input),
tag_(std::move(tag)) {
@ -194,7 +196,7 @@ class BytesProducedStatsDatasetOp : public UnaryDatasetOpKernel {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::BytesProducedStats")});
}
@ -206,7 +208,7 @@ class BytesProducedStatsDatasetOp : public UnaryDatasetOpKernel {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "BytesProducedStatsDatasetOp::Dataset";
}

View File

@ -70,7 +70,7 @@ class TakeWhileDatasetOp : public UnaryDatasetOpKernel {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::TakeWhile")});
}
@ -83,7 +83,7 @@ class TakeWhileDatasetOp : public UnaryDatasetOpKernel {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "TakeWhileDatasetOp::Dataset";
}

View File

@ -65,8 +65,8 @@ absl::Status ValidateNumThreads(int32_t num_threads) {
class ThreadPoolResource : public ResourceBase {
public:
ThreadPoolResource(Env* env, const ThreadOptions& thread_options,
const string& name, int num_threads, bool low_latency_hint,
int max_intra_op_parallelism)
const std::string& name, int num_threads,
bool low_latency_hint, int max_intra_op_parallelism)
: thread_pool_(env, thread_options, name, num_threads, low_latency_hint),
max_intra_op_parallelism_(max_intra_op_parallelism) {}
@ -86,9 +86,9 @@ class ThreadPoolResource : public ResourceBase {
}
}
int32 NumThreads() { return thread_pool_.NumThreads(); }
int32_t NumThreads() { return thread_pool_.NumThreads(); }
string DebugString() const override { return "ThreadPoolResource"; }
std::string DebugString() const override { return "ThreadPoolResource"; }
private:
thread::ThreadPool thread_pool_;
@ -156,7 +156,7 @@ class ThreadPoolHandleOp : public OpKernel {
mutex mu_;
ContainerInfo cinfo_ TF_GUARDED_BY(mu_);
bool initialized_ TF_GUARDED_BY(mu_) = false;
string display_name_;
std::string display_name_;
int num_threads_;
int max_intra_op_parallelism_;
};
@ -194,7 +194,7 @@ class ThreadPoolDatasetOp : public UnaryDatasetOpKernel {
}
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::ThreadPool")});
}
@ -206,7 +206,7 @@ class ThreadPoolDatasetOp : public UnaryDatasetOpKernel {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "ThreadPoolDatasetOp::Dataset";
}
@ -316,7 +316,7 @@ class MaxIntraOpParallelismDatasetOp::Dataset : public DatasetBase {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(Iterator::Params{
this, absl::StrCat(prefix, "::MaxIntraOpParallelism")});
}
@ -328,7 +328,7 @@ class MaxIntraOpParallelismDatasetOp::Dataset : public DatasetBase {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "MaxIntraOpParallelismDatasetOp::Dataset";
}
@ -460,7 +460,7 @@ class PrivateThreadPoolDatasetOp::Dataset : public DatasetBase {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::PrivateThreadPool")});
}
@ -472,7 +472,7 @@ class PrivateThreadPoolDatasetOp::Dataset : public DatasetBase {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return "PrivateThreadPoolDatasetOp::Dataset";
}

View File

@ -47,7 +47,7 @@ class UniqueDatasetOp::Dataset : public DatasetBase {
~Dataset() override { input_->Unref(); }
std::unique_ptr<IteratorBase> MakeIteratorInternal(
const string& prefix) const override {
const std::string& prefix) const override {
return std::make_unique<Iterator>(
Iterator::Params{this, absl::StrCat(prefix, "::Unique")});
}
@ -60,7 +60,7 @@ class UniqueDatasetOp::Dataset : public DatasetBase {
return input_->output_shapes();
}
string DebugString() const override {
std::string DebugString() const override {
return absl::StrCat("UniqueDatasetOp::Dataset");
}
@ -173,7 +173,7 @@ class UniqueDatasetOp::Dataset : public DatasetBase {
} else {
DCHECK_EQ(DT_STRING, t.dtype());
auto flat_t = t.flat<tstring>();
uint64 hash = 0;
uint64_t hash = 0;
for (int64_t i = 0; i < t.NumElements(); ++i) {
hash = Hash64Combine(hash, Hash64(flat_t(i)));
}

View File

@ -44,7 +44,8 @@ class UniqueDatasetParams : public DatasetParams {
std::vector<Tensor> GetInputTensors() const override { return {}; }
absl::Status GetInputNames(std::vector<string>* input_names) const override {
absl::Status GetInputNames(
std::vector<std::string>* input_names) const override {
input_names->clear();
input_names->emplace_back(UniqueDatasetOp::kInputDataset);
return absl::OkStatus();
@ -57,7 +58,9 @@ class UniqueDatasetParams : public DatasetParams {
return absl::OkStatus();
}
string dataset_type() const override { return UniqueDatasetOp::kDatasetType; }
std::string dataset_type() const override {
return UniqueDatasetOp::kDatasetType;
}
};
class UniqueDatasetOpTest : public DatasetOpsTestBase {};

View File

@ -168,7 +168,7 @@ class WeightedFlatMapDatasetOp::Dataset : public DatasetBase {
return output_shapes_;
}
string DebugString() const override {
std::string DebugString() const override {
return name_utils::DatasetDebugString(kDatasetType);
}