mirror of
https://github.com/lordmathis/CUDANet.git
synced 2025-11-05 17:34:21 +00:00
Update activation test
This commit is contained in:
@@ -1,6 +1,9 @@
|
|||||||
import torchvision
|
import torchvision
|
||||||
|
import torch
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
from torchsummary import summary
|
||||||
|
|
||||||
sys.path.append('../../tools') # Ugly hack
|
sys.path.append('../../tools') # Ugly hack
|
||||||
from utils import export_model_weights, print_model_parameters
|
from utils import export_model_weights, print_model_parameters
|
||||||
|
|
||||||
@@ -9,5 +12,9 @@ if __name__ == "__main__":
|
|||||||
print_model_parameters(alexnet) # print layer names and number of parameters
|
print_model_parameters(alexnet) # print layer names and number of parameters
|
||||||
export_model_weights(alexnet, 'alexnet_weights.bin')
|
export_model_weights(alexnet, 'alexnet_weights.bin')
|
||||||
print()
|
print()
|
||||||
print(alexnet)
|
|
||||||
|
if torch.cuda.is_available():
|
||||||
|
alexnet.cuda()
|
||||||
|
|
||||||
|
summary(alexnet, (3, 227, 227))
|
||||||
|
|
||||||
|
|||||||
@@ -17,6 +17,8 @@ Activation::Activation(ActivationType activation, const int length)
|
|||||||
|
|
||||||
d_softmax_sum = nullptr;
|
d_softmax_sum = nullptr;
|
||||||
CUDA_CHECK(cudaMalloc((void**)&d_softmax_sum, sizeof(float) * length));
|
CUDA_CHECK(cudaMalloc((void**)&d_softmax_sum, sizeof(float) * length));
|
||||||
|
|
||||||
|
std::cout << "Activation: Softmax " << length << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
gridSize = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
gridSize = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
||||||
|
|||||||
@@ -28,7 +28,16 @@ void Utils::max(float* d_vec, float* d_max, const unsigned int length) {
|
|||||||
|
|
||||||
const int grid_size = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
const int grid_size = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
||||||
|
|
||||||
|
std::cout << "grid_size: " << grid_size << ", length: " << length << std::endl;
|
||||||
|
CUDA_CHECK(cudaGetLastError());
|
||||||
|
|
||||||
Kernels::max_reduce<<<grid_size, BLOCK_SIZE>>>(d_vec, d_max, length);
|
Kernels::max_reduce<<<grid_size, BLOCK_SIZE>>>(d_vec, d_max, length);
|
||||||
|
|
||||||
|
std::cout << "input: " << std::endl;
|
||||||
|
print_vec(d_vec, length);
|
||||||
|
std::cout << "max: " << std::endl;
|
||||||
|
print_vec(d_max, length);
|
||||||
|
|
||||||
CUDA_CHECK(cudaGetLastError());
|
CUDA_CHECK(cudaGetLastError());
|
||||||
|
|
||||||
int remaining = grid_size;
|
int remaining = grid_size;
|
||||||
@@ -46,7 +55,6 @@ void Utils::sum(float* d_vec, float* d_sum, const unsigned int length) {
|
|||||||
|
|
||||||
const int gridSize = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
const int gridSize = (length + BLOCK_SIZE - 1) / BLOCK_SIZE;
|
||||||
|
|
||||||
|
|
||||||
CUDANet::Kernels::sum_reduce<<<gridSize, BLOCK_SIZE>>>(
|
CUDANet::Kernels::sum_reduce<<<gridSize, BLOCK_SIZE>>>(
|
||||||
d_vec, d_sum, length
|
d_vec, d_sum, length
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -4,26 +4,33 @@
|
|||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
TEST(ActivationTest, SoftmaxTest1) {
|
TEST(ActivationTest, SoftmaxTest1) {
|
||||||
|
const int inputSize = 5;
|
||||||
|
cudaError_t cudaStatus;
|
||||||
|
|
||||||
CUDANet::Layers::Activation activation(
|
CUDANet::Layers::Activation activation(
|
||||||
CUDANet::Layers::ActivationType::SOFTMAX, 5
|
CUDANet::Layers::ActivationType::SOFTMAX, inputSize
|
||||||
);
|
);
|
||||||
|
|
||||||
std::vector<float> input = {0.573f, 0.619f, 0.732f, 0.055f, 0.243f};
|
std::vector<float> input = {0.573f, 0.619f, 0.732f, 0.055f, 0.243f};
|
||||||
|
|
||||||
float* d_input;
|
float* d_input;
|
||||||
cudaMalloc((void**)&d_input, sizeof(float) * 5);
|
cudaStatus = cudaMalloc((void**)&d_input, sizeof(float) * inputSize);
|
||||||
cudaMemcpy(d_input, input.data(), sizeof(float) * 5, cudaMemcpyHostToDevice);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
|
cudaStatus = cudaMemcpy(d_input, input.data(), sizeof(float) * inputSize, cudaMemcpyHostToDevice);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
activation.activate(d_input);
|
activation.activate(d_input);
|
||||||
std::vector<float> output(5);
|
std::vector<float> output(5);
|
||||||
cudaMemcpy(
|
cudaStatus = cudaMemcpy(
|
||||||
output.data(), d_input, sizeof(float) * 5, cudaMemcpyDeviceToHost
|
output.data(), d_input, sizeof(float) * inputSize, cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
float sum = 0.0f;
|
float sum = 0.0f;
|
||||||
|
|
||||||
std::vector<float> expected = {0.22055f, 0.23094f, 0.25856f, 0.13139f, 0.15856f};
|
std::vector<float> expected = {0.22055f, 0.23094f, 0.25856f, 0.13139f, 0.15856f};
|
||||||
for (int i = 0; i < 5; ++i) {
|
for (int i = 0; i < inputSize; ++i) {
|
||||||
sum += output[i];
|
sum += output[i];
|
||||||
EXPECT_NEAR(output[i], expected[i], 1e-5f);
|
EXPECT_NEAR(output[i], expected[i], 1e-5f);
|
||||||
}
|
}
|
||||||
@@ -35,32 +42,42 @@ TEST(ActivationTest, SoftmaxTest1) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
TEST(ActivationTest, SoftmaxTest2) {
|
TEST(ActivationTest, SoftmaxTest2) {
|
||||||
|
const int inputSize = 6;
|
||||||
|
cudaError_t cudaStatus;
|
||||||
|
|
||||||
CUDANet::Layers::Activation activation(
|
CUDANet::Layers::Activation activation(
|
||||||
CUDANet::Layers::ActivationType::SOFTMAX, 6
|
CUDANet::Layers::ActivationType::SOFTMAX, inputSize
|
||||||
);
|
);
|
||||||
|
|
||||||
|
cudaStatus = cudaGetLastError();
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
std::vector<float> input = {22.496f, 36.9006f, 30.9904f, 28.4213f, 26.4541f, 31.7887f};
|
std::vector<float> input = {22.496f, 36.9006f, 30.9904f, 28.4213f, 26.4541f, 31.7887f};
|
||||||
|
|
||||||
float* d_input;
|
float* d_input;
|
||||||
cudaMalloc((void**)&d_input, sizeof(float) * 6);
|
cudaStatus = cudaMalloc((void**)&d_input, sizeof(float) * inputSize);
|
||||||
cudaMemcpy(d_input, input.data(), sizeof(float) * 6, cudaMemcpyHostToDevice);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
|
cudaStatus = cudaMemcpy(d_input, input.data(), sizeof(float) * inputSize, cudaMemcpyHostToDevice);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
activation.activate(d_input);
|
activation.activate(d_input);
|
||||||
std::vector<float> output(6);
|
std::vector<float> output(inputSize);
|
||||||
cudaMemcpy(
|
cudaStatus = cudaMemcpy(
|
||||||
output.data(), d_input, sizeof(float) * 6, cudaMemcpyDeviceToHost
|
output.data(), d_input, sizeof(float) * inputSize, cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
float sum = 0.0f;
|
float sum = 0.0f;
|
||||||
|
|
||||||
std::vector<float> expected = {0.0f, 0.99111f, 0.00269f, 0.00021f, 3e-05f, 0.00597f};
|
std::vector<float> expected = {0.0f, 0.99111f, 0.00269f, 0.00021f, 3e-05f, 0.00597f};
|
||||||
for (int i = 0; i < 5; ++i) {
|
for (int i = 0; i < inputSize; ++i) {
|
||||||
sum += output[i];
|
sum += output[i];
|
||||||
EXPECT_NEAR(output[i], expected[i], 1e-5f);
|
EXPECT_NEAR(output[i], expected[i], 1e-5f);
|
||||||
}
|
}
|
||||||
|
EXPECT_NEAR(sum, 1.0f, 1e-5f);
|
||||||
|
|
||||||
EXPECT_NEAR(sum, 1.0f, 1e-2f);
|
// Cleanup
|
||||||
|
|
||||||
cudaFree(d_input);
|
cudaFree(d_input);
|
||||||
cudaDeviceReset();
|
cudaDeviceReset();
|
||||||
}
|
}
|
||||||
Reference in New Issue
Block a user