From a40f3bdebd5971d4e680e05237cb987716d82356 Mon Sep 17 00:00:00 2001 From: Bill Varcho Date: Mon, 20 Oct 2025 15:10:39 -0700 Subject: [PATCH] [Upkeep] Resolve 4 instances of the following issue: Todo (resolved) PiperOrigin-RevId: 821824006 --- .../transforms/collectives/collectives_schedule_linearizer.cc | 1 - .../xla/xla/tests/collective_pipeline_parallelism_test.cc | 2 -- third_party/xla/xla/tools/collective_perf_table_gen_main.cc | 2 -- 3 files changed, 5 deletions(-) diff --git a/third_party/xla/xla/hlo/transforms/collectives/collectives_schedule_linearizer.cc b/third_party/xla/xla/hlo/transforms/collectives/collectives_schedule_linearizer.cc index 2508100d6c8..495abf85385 100644 --- a/third_party/xla/xla/hlo/transforms/collectives/collectives_schedule_linearizer.cc +++ b/third_party/xla/xla/hlo/transforms/collectives/collectives_schedule_linearizer.cc @@ -31,7 +31,6 @@ limitations under the License. namespace xla { -// TODO(b/181653482): Fix for interprocedural collectives as well. absl::StatusOr CollectivesScheduleLinearizer::Run( HloModule* module, const absl::flat_hash_set& execution_threads) { diff --git a/third_party/xla/xla/tests/collective_pipeline_parallelism_test.cc b/third_party/xla/xla/tests/collective_pipeline_parallelism_test.cc index 6e89f6232bd..afdc58a81b7 100644 --- a/third_party/xla/xla/tests/collective_pipeline_parallelism_test.cc +++ b/third_party/xla/xla/tests/collective_pipeline_parallelism_test.cc @@ -1274,7 +1274,6 @@ TEST_P(CollectivePipelineParallelismTest, // This is the partially pipelined version of // NaiveBFSMicrobatch5CircularRepeat2Replica4 and should yield the same results. -// TODO(b/383868854): replace this with GPU pipeliner implementation. TEST_P(CollectivePipelineParallelismTest, NaiveBFSMb5Cr2Replica4SendRecvPartiallyPipelined) { constexpr char kMoreComputationsStr[] = R"( @@ -1508,7 +1507,6 @@ TEST_P(CollectivePipelineParallelismTest, // This is the async-grouped version of // NaiveBFSMicrobatch5CircularRepeat2Replica4 and should yield the same results. -// TODO(b/383868854): replace this with GPU pipeliner implementation. TEST_P(CollectivePipelineParallelismTest, NaiveBFSMb5Cr2Replica4SendRecvAsyncGroup) { constexpr char kMoreComputationsStr[] = R"( diff --git a/third_party/xla/xla/tools/collective_perf_table_gen_main.cc b/third_party/xla/xla/tools/collective_perf_table_gen_main.cc index 51d3a5dfb54..cf581896f88 100644 --- a/third_party/xla/xla/tools/collective_perf_table_gen_main.cc +++ b/third_party/xla/xla/tools/collective_perf_table_gen_main.cc @@ -158,8 +158,6 @@ std::string DefaultCollectiveDevicesIfEmpty( } // namespace -// TODO(b/390097558): Add an option to generate perf table for collective which -// gets overlap to model resource contention. int main(int argc, char* argv[]) { // Default args. int32_t num_nodes = 1;