|
|
4a1c4a5f91
|
Add epsilon param to batch norm
|
2024-05-19 15:13:22 +02:00 |
|
|
|
33a3c6afcb
|
Fix batch norm layer
|
2024-05-14 21:42:01 +02:00 |
|
|
|
da8f3167cb
|
Add utils vector mean function
|
2024-05-14 21:41:18 +02:00 |
|
|
|
96536a44fd
|
Implement batch norm test
|
2024-04-28 21:46:42 +02:00 |
|
|
|
69111f6cb1
|
Add version to model bin format
|
2024-04-23 19:53:48 +02:00 |
|
|
|
f17debc244
|
Implement getOutputSize and getInputSize for seq layers
|
2024-04-22 20:31:58 +02:00 |
|
|
|
fbf6c44bdd
|
Remove extra cudaFree
|
2024-04-22 19:05:02 +02:00 |
|
|
|
26cea9b12c
|
Fix more cuda memcheck errors
|
2024-04-21 22:56:55 +02:00 |
|
|
|
58af95eb25
|
Remove cudaDeviceReset calls from tests
|
2024-04-21 22:47:12 +02:00 |
|
|
|
bdbb3f978e
|
Fix matmul and max reduce memcheck errors
|
2024-04-21 22:11:02 +02:00 |
|
|
|
9a6152469a
|
Update activation test
|
2024-04-21 14:00:43 +02:00 |
|
|
|
d64a28bc9c
|
Fix model weights export
|
2024-04-21 00:05:56 +02:00 |
|
|
|
ecf7416f8e
|
Rework padding size setting
|
2024-04-20 16:31:28 +02:00 |
|
|
|
432adf57bd
|
Test model weights loading
|
2024-04-16 21:07:06 +02:00 |
|
|
|
f4ae45f867
|
Start implementing weights import
|
2024-04-15 22:17:48 +02:00 |
|
|
|
d8c50116e8
|
Implement pytorch weights export
|
2024-04-15 22:17:14 +02:00 |
|
|
|
13b455e4ba
|
Implement test multiple inference
|
2024-04-14 00:19:29 +02:00 |
|
|
|
53c976733b
|
Refactor model test
|
2024-04-14 00:05:32 +02:00 |
|
|
|
457e0f669f
|
Implement model test
|
2024-04-13 13:31:29 +02:00 |
|
|
|
18522c2dea
|
Cleanup and refactor
|
2024-04-11 22:52:41 +02:00 |
|
|
|
4b9d123e94
|
Implement device vector utils
|
2024-04-11 22:22:33 +02:00 |
|
|
|
710a33bdde
|
Move softmax partial kernels to matmul
|
2024-04-11 22:01:47 +02:00 |
|
|
|
bf7c961b9e
|
Add cudaDeviceReset at the end of each test
|
2024-04-11 19:55:02 +02:00 |
|
|
|
b49dddf34a
|
Improve softmax numerical stability
|
2024-04-08 23:25:46 +02:00 |
|
|
|
e419a93408
|
Fix softmax sum kernel
|
2024-04-08 22:09:18 +02:00 |
|
|
|
7bc329a043
|
Add more softmax tests
|
2024-03-22 22:32:08 +01:00 |
|
|
|
9482d7bc43
|
Add model predict test
|
2024-03-22 22:31:32 +01:00 |
|
|
|
90fb104dae
|
Implement output layer
|
2024-03-21 23:07:46 +01:00 |
|
|
|
6f4cdf3792
|
Implement avg pool test
|
2024-03-20 21:57:22 +01:00 |
|
|
|
dfff0360d9
|
Implement max pooling test
|
2024-03-20 21:44:04 +01:00 |
|
|
|
6cf604423a
|
Combine padding and conv kernel
|
2024-03-18 19:53:40 +01:00 |
|
|
|
aac0c3a826
|
Implement concat layer
|
2024-03-17 21:38:29 +01:00 |
|
|
|
cbdb4e7707
|
Test softmax
|
2024-03-17 19:08:16 +01:00 |
|
|
|
42d646750b
|
Abstract activation and implement softmax
|
2024-03-17 18:37:15 +01:00 |
|
|
|
0c22fac64e
|
Add toplevel CUDANet namespace
|
2024-03-17 16:08:53 +01:00 |
|
|
|
77004c16be
|
Use shared memory for mat vec mul kernel
|
2024-03-13 22:13:11 +01:00 |
|
|
|
7157a27e56
|
Add documentation comments
|
2024-03-12 21:50:06 +01:00 |
|
|
|
708164e4d0
|
Implement simple input layer
|
2024-03-12 21:16:46 +01:00 |
|
|
|
9d91896f13
|
Change forward function to return output pointer
|
2024-03-12 20:50:49 +01:00 |
|
|
|
2518138ef8
|
Add strided conv2d test
|
2024-03-11 21:51:27 +01:00 |
|
|
|
d2ab78fbc7
|
Add Kernels namespace
|
2024-03-11 21:04:23 +01:00 |
|
|
|
e0178e2d5c
|
Cleanup and refactor
|
2024-03-11 20:39:44 +01:00 |
|
|
|
f3112311da
|
Make conv2d work again
|
2024-03-10 19:13:22 +01:00 |
|
|
|
96804777ee
|
Refactor conv2d test
|
2024-03-09 22:54:46 +01:00 |
|
|
|
a3d85a10fc
|
Working conv2d forward
|
2024-03-09 21:08:16 +01:00 |
|
|
|
4b6fcbc191
|
Implement simple test for host conv2d
|
2024-03-08 23:12:04 +01:00 |
|
|
|
69ccba2dad
|
Start conv test implementation
|
2024-03-07 22:03:05 +01:00 |
|
|
|
07f231a30b
|
Switch padding kernel to row major
|
2024-03-05 21:04:11 +01:00 |
|
|
|
f4257afd5a
|
Remove cublas dependency
|
2024-03-05 18:41:35 +01:00 |
|
|
|
f37320594a
|
Add activations enum
|
2024-03-03 15:24:54 +01:00 |
|