223 Commits

Author SHA1 Message Date
364715ff70 Refactor kernels 2024-03-19 21:37:25 +01:00
b6c4b7d2ae Refactor layers 2024-03-19 21:35:05 +01:00
8d14b74f66 Implement Add layer 2024-03-18 20:37:13 +01:00
d9c6c663c8 Rename ILayer to WeightedLayer 2024-03-18 20:36:52 +01:00
6cf604423a Combine padding and conv kernel 2024-03-18 19:53:40 +01:00
e6d3757312 Change unsigned int to int 2024-03-18 19:40:00 +01:00
aac0c3a826 Implement concat layer 2024-03-17 21:38:29 +01:00
9aab7cdb10 Update README 2024-03-17 19:12:39 +01:00
48060dc38d Rename python tool script 2024-03-17 19:11:41 +01:00
cbdb4e7707 Test softmax 2024-03-17 19:08:16 +01:00
42d646750b Abstract activation and implement softmax 2024-03-17 18:37:15 +01:00
b1621819ca Update README 2024-03-17 16:18:42 +01:00
0c22fac64e Add toplevel CUDANet namespace 2024-03-17 16:08:53 +01:00
dc86cddeb7 Use tiling shmem for mat vec mul kernel 2024-03-15 23:33:09 +01:00
88f7fff217 Add prefix to guards 2024-03-13 22:23:23 +01:00
77004c16be Use shared memory for mat vec mul kernel 2024-03-13 22:13:11 +01:00
09480e42e5 Add doxygen config 2024-03-12 22:09:37 +01:00
7157a27e56 Add documentation comments 2024-03-12 21:50:06 +01:00
708164e4d0 Implement simple input layer 2024-03-12 21:16:46 +01:00
9d91896f13 Change forward function to return output pointer 2024-03-12 20:50:49 +01:00
2518138ef8 Add strided conv2d test 2024-03-11 21:51:27 +01:00
a3973f0b21 Add activation to conv2d 2024-03-11 21:05:38 +01:00
d2ab78fbc7 Add Kernels namespace 2024-03-11 21:04:23 +01:00
e0178e2d5c Cleanup and refactor 2024-03-11 20:39:44 +01:00
f3112311da Make conv2d work again 2024-03-10 19:13:22 +01:00
6bbc036f62 Generate conv2d test results with pytorch 2024-03-10 19:11:48 +01:00
6ce45cc834 Remove gh actions lint 2024-03-09 23:21:36 +01:00
d177a67cd6 Add bias to conv2d 2024-03-09 23:03:23 +01:00
4f3c4f1afb Fix conv2d kernel dims 2024-03-09 22:55:37 +01:00
96804777ee Refactor conv2d test 2024-03-09 22:54:46 +01:00
fceef07a9b Remove linear activation kernel 2024-03-09 22:54:23 +01:00
a3d85a10fc Working conv2d forward 2024-03-09 21:08:16 +01:00
e51aabc2f2 Initial cuda conv kernel implementation 2024-03-08 23:35:54 +01:00
4b6fcbc191 Implement simple test for host conv2d 2024-03-08 23:12:04 +01:00
69ccba2dad Start conv test implementation 2024-03-07 22:03:05 +01:00
7e75943a6b Add stride to index calculation 2024-03-07 21:32:39 +01:00
fc2c1616b4 Initial cpu conv implementation 2024-03-07 21:24:59 +01:00
07f231a30b Switch padding kernel to row major 2024-03-05 21:04:11 +01:00
f4257afd5a Remove cublas dependency 2024-03-05 18:41:35 +01:00
98ad84c659 Add matrix math kernels 2024-03-05 17:38:46 +01:00
cfc5c46d5e Initialize conv2d layer 2024-03-04 22:16:03 +01:00
f37320594a Add activations enum 2024-03-03 15:24:54 +01:00
7e4460cc5e Implement padding kernel 2024-03-03 15:24:39 +01:00
019ccc33d9 Start implementing padding kernel 2024-02-29 22:21:48 +01:00
045359cca2 Remove not needed code 2024-02-29 22:21:32 +01:00
0f0e57b819 Update ReLU test 2024-02-28 21:45:52 +01:00
e267f08a2f Format code 2024-02-27 21:49:05 +01:00
19ee20ea66 Add dense sigmoid test 2024-02-27 21:48:08 +01:00
9747abe53e Fix dense layer forward prop 2024-02-27 21:47:46 +01:00
b1eb8b5806 Add activations test 2024-02-27 20:19:17 +01:00