Clean up verbose logging in TPUPartitionedCallOp.

PiperOrigin-RevId: 365702660
Change-Id: Ifba977b2e272cb1a2c4616d9d49b9d5cfdb36e65
diff --git a/tensorflow/core/tpu/kernels/tpu_functional_ops.cc b/tensorflow/core/tpu/kernels/tpu_functional_ops.cc
index 7af3bdd..b5e10ae 100644
--- a/tensorflow/core/tpu/kernels/tpu_functional_ops.cc
+++ b/tensorflow/core/tpu/kernels/tpu_functional_ops.cc
@@ -1868,8 +1868,8 @@
         graph, xla_sharding_ops, tpu_replicated_input_ops);
   }
 
-  VLOG(3) << "xla_spmd_input_sharded: " << xla_spmd_input_sharded;
-  VLOG(1) << DumpGraphToFile("before_get_input_output_info1", *graph,
+  VLOG(1) << "xla_spmd_input_sharded: " << xla_spmd_input_sharded;
+  VLOG(2) << DumpGraphToFile("before_remove_descendant_nodes", *graph,
                              flib_def_.get());
   if (xla_spmd_input_sharded) {
     // We are setting must_be_child_of_arg == true because we do not want
@@ -1888,14 +1888,14 @@
                               /*must_be_child_of_arg=*/false);
   }
 
-  VLOG(1) << DumpGraphToFile("before_get_input_output_info2", *graph,
+  VLOG(2) << DumpGraphToFile("before_get_input_output_info", *graph,
                              flib_def_.get());
 
   TF_RETURN_IF_ERROR(GetInputOutputInfo(graph, tpu_inferred_info, arg_shapes,
                                         tpu_input_shapes, tpu_input_dtypes,
                                         ctx));
 
-  VLOG(1) << DumpGraphToFile("before_optimize_tpu_input_output_tensors", *graph,
+  VLOG(2) << DumpGraphToFile("before_optimize_tpu_input_output_tensors", *graph,
                              flib_def_.get());
 
   string cluster_name;