18 | const int tid = threadIdx.x;
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:19:21: warning: narrowing conversion from 'unsigned int' to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
19 | const int gid = blockIdx.x * blockDim.x + tid;
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:20:29: warning: narrowing conversion from 'unsigned int' to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
20 | const int grid_stride = gridDim.x * blockDim.x;
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:47:5: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
47 | if (tid < 256) { shm[tid] += shm[tid + 256]; } __syncthreads();
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:49:5: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
49 | if (tid < 128) { shm[tid] += shm[tid + 128]; } __syncthreads();
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:51:5: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
51 | if (tid < 64) { shm[tid] += shm[tid + 64]; } __syncthreads();
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:57:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
57 | if (tid < 32) vmem[tid] += vmem[tid + 32];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:59:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
59 | if (tid < 16) vmem[tid] += vmem[tid + 16];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:61:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
61 | if (tid < 8) vmem[tid] += vmem[tid + 8];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:63:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
63 | if (tid < 4) vmem[tid] += vmem[tid + 4];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:65:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
65 | if (tid < 2) vmem[tid] += vmem[tid + 2];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:67:9: error: expected a for, while, or do-while loop to follow '#pragma unroll' [clang-diagnostic-error]
67 | if (tid < 1) vmem[tid] += vmem[tid + 1];
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250212_optimize_b5_s4_e1_v2/level_1/task_94/b4_s2_mse_unrolled_optimized/edit_1/edit_1.cu:91:5: warning: inside a lambda, '__func__' expands to the name of the function call operator; consider capturing the name of the enclosing function explicitly [bugprone-lambda-function-name]
91 | AT_DISPATCH_FLOATING_TYPES(predictions.scalar_type(), "mse_forward_cuda", [&] {
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:237:34: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES'
237 | AT_DISPATCH_SWITCH(TYPE, NAME, AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__))
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:233:3: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES'
233 | AT_DISPATCH_CASE(at::ScalarType::Double, __VA_ARGS__) \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:74:3: note: expanded from macro 'AT_DISPATCH_CASE'
74 | AT_PRIVATE_CASE_TYPE_USING_HINT(enum_type, scalar_t, __VA_ARGS__)
| ^
note: (skipping 1 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all)
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:58:7: note: expanded from macro 'AT_PRIVATE_CHECK_SELECTIVE_BUILD'
58 | AT_ERROR( \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/c10/util/Exception.h:711:32: note: expanded from macro 'AT_ERROR'
711 | C10_EXPAND_MSVC_WORKAROUND(TORCH_CHECK(false, ::c10::str(__VA_ARGS__))); \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/c10/util/Exception.h:536:9: note: expanded from macro 'TORCH_CHECK'
536 | __func__, \
| ^