12 | const scalar_t* __restrict__ weight,
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
13 | const scalar_t* __restrict__ bias,
| ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:12:34: note: the first parameter in the range is 'weight'
12 | const scalar_t* __restrict__ weight,
| ^~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:13:34: note: the last parameter in the range is 'bias'
13 | const scalar_t* __restrict__ bias,
| ^~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:15:5: warning: 2 adjacent parameters of 'conv_transpose2d_kernel' of similar type ('const int') are easily swapped by mistake [bugprone-easily-swappable-parameters]
15 | const int batch_size,
| ^~~~~~~~~~~~~~~~~~~~~
16 | const int in_channels,
| ~~~~~~~~~~~~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:15:15: note: the first parameter in the range is 'batch_size'
15 | const int batch_size,
| ^~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:16:15: note: the last parameter in the range is 'in_channels'
16 | const int in_channels,
| ^~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:18:5: warning: 3 adjacent parameters of 'conv_transpose2d_kernel' of similar type ('const int') are easily swapped by mistake [bugprone-easily-swappable-parameters]
18 | const int in_width,
| ^~~~~~~~~~~~~~~~~~~
19 | const int out_channels,
| ~~~~~~~~~~~~~~~~~~~~~~~
20 | const int kernel_h,
| ~~~~~~~~~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:18:15: note: the first parameter in the range is 'in_width'
18 | const int in_width,
| ^~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:20:15: note: the last parameter in the range is 'kernel_h'
20 | const int kernel_h,
| ^~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:21:5: warning: 5 adjacent parameters of 'conv_transpose2d_kernel' of similar type ('const int') are easily swapped by mistake [bugprone-easily-swappable-parameters]
21 | const int kernel_w,
| ^~~~~~~~~~~~~~~~~~~
22 | const int stride,
| ~~~~~~~~~~~~~~~~~
23 | const int padding,
| ~~~~~~~~~~~~~~~~~~
24 | const int output_padding,
| ~~~~~~~~~~~~~~~~~~~~~~~~~
25 | const int groups,
| ~~~~~~~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:21:15: note: the first parameter in the range is 'kernel_w'
21 | const int kernel_w,
| ^~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:25:15: note: the last parameter in the range is 'groups'
25 | const int groups,
| ^~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:26:5: warning: 2 adjacent parameters of 'conv_transpose2d_kernel' of similar type ('const int') are easily swapped by mistake [bugprone-easily-swappable-parameters]
26 | const int dilation,
| ^~~~~~~~~~~~~~~~~~~
27 | const int out_height,
| ~~~~~~~~~~~~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:26:15: note: the first parameter in the range is 'dilation'
26 | const int dilation,
| ^~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:27:15: note: the last parameter in the range is 'out_height'
27 | const int out_height,
| ^~~~~~~~~~
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:35:21: warning: narrowing conversion from 'unsigned int' to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
35 | const int idx = blockIdx.x * blockDim.x + threadIdx.x;
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:107:28: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
107 | const int batch_size = x.size(0);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:108:29: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
108 | const int in_channels = x.size(1);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:109:27: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
109 | const int in_height = x.size(2);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:110:26: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
110 | const int in_width = x.size(3);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:113:30: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
113 | const int out_channels = weight.size(1) * groups;
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:114:26: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
114 | const int kernel_h = weight.size(2);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:115:26: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
115 | const int kernel_w = weight.size(3);
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:127:32: warning: narrowing conversion from 'int64_t' (aka 'long') to signed type 'int' is implementation-defined [bugprone-narrowing-conversions]
127 | const int total_elements = output.numel();
| ^
/home/robert_sakana_ai/llm_cuda/experiments/20250207_optimize_b5_s4_e1_sweep/level_1/task_65/b3_s0_balanced_workload_distribution/edit_1/edit_1.cu:131:5: warning: inside a lambda, '__func__' expands to the name of the function call operator; consider capturing the name of the enclosing function explicitly [bugprone-lambda-function-name]
131 | AT_DISPATCH_FLOATING_TYPES(x.scalar_type(), "conv_transpose2d_cuda", ([&] {
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:237:34: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES'
237 | AT_DISPATCH_SWITCH(TYPE, NAME, AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__))
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:233:3: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES'
233 | AT_DISPATCH_CASE(at::ScalarType::Double, __VA_ARGS__) \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:74:3: note: expanded from macro 'AT_DISPATCH_CASE'
74 | AT_PRIVATE_CASE_TYPE_USING_HINT(enum_type, scalar_t, __VA_ARGS__)
| ^
note: (skipping 1 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all)
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/ATen/Dispatch.h:58:7: note: expanded from macro 'AT_PRIVATE_CHECK_SELECTIVE_BUILD'
58 | AT_ERROR( \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/c10/util/Exception.h:711:32: note: expanded from macro 'AT_ERROR'
711 | C10_EXPAND_MSVC_WORKAROUND(TORCH_CHECK(false, ::c10::str(__VA_ARGS__))); \
| ^
/home/robert_sakana_ai/miniconda3/envs/llm2cuda/lib/python3.11/site-packages/torch/include/c10/util/Exception.h:536:9: note: expanded from macro 'TORCH_CHECK'
536 | __func__, \
| ^