diff --git a/backends/vulkan/runtime/api/Adapter.cpp b/backends/vulkan/runtime/api/Adapter.cpp index 3d5c87d23d..b1930ad1de 100644 --- a/backends/vulkan/runtime/api/Adapter.cpp +++ b/backends/vulkan/runtime/api/Adapter.cpp @@ -401,8 +401,7 @@ std::string Adapter::stringize() const { ss << " Memory Info {" << std::endl; ss << " Memory Types [" << std::endl; for (size_t i = 0; i < mem_props.memoryTypeCount; ++i) { - ss << " " - << " [Heap " << mem_props.memoryTypes[i].heapIndex << "] " + ss << " " << " [Heap " << mem_props.memoryTypes[i].heapIndex << "] " << get_memory_properties_str(mem_props.memoryTypes[i].propertyFlags) << std::endl; } diff --git a/backends/vulkan/runtime/graph/ops/glsl/indexing_utils.h b/backends/vulkan/runtime/graph/ops/glsl/indexing_utils.h index a5ed5b5f18..25a8060203 100644 --- a/backends/vulkan/runtime/graph/ops/glsl/indexing_utils.h +++ b/backends/vulkan/runtime/graph/ops/glsl/indexing_utils.h @@ -50,8 +50,8 @@ // describe sizes. As an example, let's say we want to swap dimensions 0,1 for a // tensor of shape {4,3,2,24} to obtain {3,4,2,24}. Then, x=4, y=3 and // plane=2*24=48. -#define SWAP_ADJ_DIMS(cur, x, y, plane) \ - cur + \ - plane*( \ - (1 - y) * ((cur % (x * y * plane)) / (y * plane)) + \ - (x - 1) * ((cur % (y * plane)) / plane)) +#define SWAP_ADJ_DIMS(cur, x, y, plane) \ + cur + \ + plane * \ + ((1 - y) * ((cur % (x * y * plane)) / (y * plane)) + \ + (x - 1) * ((cur % (y * plane)) / plane)) diff --git a/examples/models/llama2/runner/runner.cpp b/examples/models/llama2/runner/runner.cpp index 2677e6ad80..61a5ea66bd 100644 --- a/examples/models/llama2/runner/runner.cpp +++ b/examples/models/llama2/runner/runner.cpp @@ -472,8 +472,7 @@ std::string statsToJsonString(const Runner::Stats& stats) { << "\"prompt_eval_end_ms\":" << stats.prompt_eval_end_ms << "," << "\"first_token_ms\":" << stats.first_token_ms << "," << "\"aggregate_sampling_time_ms\":" << stats.aggregate_sampling_time_ms - << "," - << "\"SCALING_FACTOR_UNITS_PER_SECOND\":" + << "," << "\"SCALING_FACTOR_UNITS_PER_SECOND\":" << stats.SCALING_FACTOR_UNITS_PER_SECOND << "}"; return ss.str(); } diff --git a/kernels/portable/cpu/op_cumsum.cpp b/kernels/portable/cpu/op_cumsum.cpp index 7fe483ad2e..fffc2d4639 100644 --- a/kernels/portable/cpu/op_cumsum.cpp +++ b/kernels/portable/cpu/op_cumsum.cpp @@ -11,8 +11,8 @@ #include #include #include -//#include -//#include +// #include +// #include namespace torch { namespace executor { diff --git a/runtime/core/portable_type/optional.h b/runtime/core/portable_type/optional.h index 94c0caa8a2..3d8cb41eac 100644 --- a/runtime/core/portable_type/optional.h +++ b/runtime/core/portable_type/optional.h @@ -74,8 +74,8 @@ class optional final { } optional& operator=(optional&& rhs) noexcept( - std::is_nothrow_move_assignable::value&& - std::is_nothrow_move_constructible::value) { + std::is_nothrow_move_assignable::value && + std::is_nothrow_move_constructible::value) { if (init_ && !rhs.init_) { clear(); } else if (!init_ && rhs.init_) { diff --git a/sdk/etdump/etdump_flatcc.cpp b/sdk/etdump/etdump_flatcc.cpp index 4e67532ca8..da4851214f 100644 --- a/sdk/etdump/etdump_flatcc.cpp +++ b/sdk/etdump/etdump_flatcc.cpp @@ -103,7 +103,8 @@ ETDumpGen::ETDumpGen(Span buffer) { alloc.set_buffer( (uint8_t*)buffer_with_builder, buffer_size, - (size_t)((buffer_size / 4 > max_alloc_buf_size) ? max_alloc_buf_size : buffer_size / 4)); + (size_t)((buffer_size / 4 > max_alloc_buf_size) ? max_alloc_buf_size + : buffer_size / 4)); et_flatcc_custom_init(builder, &alloc); } else { builder = (struct flatcc_builder*)malloc(sizeof(struct flatcc_builder));