Fix sanity checks
diff --git a/tensorflow/cc/saved_model/BUILD b/tensorflow/cc/saved_model/BUILD
index c48f898..eea5f47 100644
--- a/tensorflow/cc/saved_model/BUILD
+++ b/tensorflow/cc/saved_model/BUILD
@@ -45,7 +45,7 @@
     deps = [
         ":constants",
         "//tensorflow/core:protos_all_cc",
-       "//tensorflow/core/util/tensor_bundle",
+        "//tensorflow/core/util/tensor_bundle",
     ] + if_not_mobile([
         # TODO(b/111634734): :lib and :protos_all contain dependencies that
         # cannot be built on mobile platforms. Instead, include the appropriate
diff --git a/tensorflow/python/saved_model/load.py b/tensorflow/python/saved_model/load.py
index c4eb751..e23be2e 100644
--- a/tensorflow/python/saved_model/load.py
+++ b/tensorflow/python/saved_model/load.py
@@ -878,7 +878,8 @@
     # tensor_content field contains raw bytes in litle endian format which causes problems
     # when loaded on big-endian systems requiring byteswap
     if sys.byteorder == 'big':
-      saved_model_utils.swap_function_tensor_content(meta_graph_def, "little", "big")
+      saved_model_utils.swap_function_tensor_content(
+              meta_graph_def, "little", "big")
     if (tags is not None
         and set(tags) != set(meta_graph_def.meta_info_def.tags)):
       raise ValueError(
diff --git a/tensorflow/python/saved_model/utils_impl.py b/tensorflow/python/saved_model/utils_impl.py
index 70135e3..19c1dff 100644
--- a/tensorflow/python/saved_model/utils_impl.py
+++ b/tensorflow/python/saved_model/utils_impl.py
@@ -318,6 +318,9 @@
         tensor_size = tensor_size*sz.size
       chunksize = int(len(tensor_bytes)/tensor_size)
       #split tensor_data into chunks for byte swapping
-      to_swap = [tensor_bytes[i:i+chunksize] for i in range(0, len(tensor_bytes), chunksize)]
+      to_swap = [tensor_bytes[i:i+chunksize] for i in range(
+          0, len(tensor_bytes), chunksize)]
       #swap and replace tensor_content
-      tensor.tensor_content = b''.join([int.from_bytes(byteswap, from_endiness).to_bytes(chunksize, to_endiness) for byteswap in to_swap])
+      tensor.tensor_content = b''.join([int.from_bytes(
+          byteswap, from_endiness).to_bytes(
+              chunksize, to_endiness) for byteswap in to_swap])