mirror of
https://github.com/lordmathis/CUDANet.git
synced 2025-11-06 01:34:22 +00:00
Remove cublas dependency
This commit is contained in:
@@ -12,6 +12,7 @@ set(LIBRARY_SOURCES
|
|||||||
src/utils/cuda_helper.cu
|
src/utils/cuda_helper.cu
|
||||||
src/kernels/activations.cu
|
src/kernels/activations.cu
|
||||||
src/kernels/padding.cu
|
src/kernels/padding.cu
|
||||||
|
src/kernels/matrix_math.cu
|
||||||
src/layers/dense.cu
|
src/layers/dense.cu
|
||||||
src/layers/conv2d.cu
|
src/layers/conv2d.cu
|
||||||
)
|
)
|
||||||
@@ -23,8 +24,7 @@ set(CUDA_NVCC_FLAGS ${CUDA_NVCC_FLAGS} -arch=sm_75)
|
|||||||
# Build static library
|
# Build static library
|
||||||
add_library(${PROJECT_NAME} STATIC ${LIBRARY_SOURCES})
|
add_library(${PROJECT_NAME} STATIC ${LIBRARY_SOURCES})
|
||||||
|
|
||||||
# Link cuBLAS library to the library
|
target_link_libraries(${PROJECT_NAME} CUDA::cudart)
|
||||||
target_link_libraries(${PROJECT_NAME} CUDA::cublas CUDA::cudart)
|
|
||||||
|
|
||||||
# Set include directories for the library
|
# Set include directories for the library
|
||||||
target_include_directories(${PROJECT_NAME} PUBLIC
|
target_include_directories(${PROJECT_NAME} PUBLIC
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
# CUDANet
|
# CUDANet
|
||||||
|
|
||||||
requirements:
|
requirements:
|
||||||
- CUDA, cuBLAS
|
- CUDA
|
||||||
- Google Test
|
- Google Test
|
||||||
@@ -1,8 +1,6 @@
|
|||||||
#ifndef CONV_LAYER_H
|
#ifndef CONV_LAYER_H
|
||||||
#define CONV_LAYER_H
|
#define CONV_LAYER_H
|
||||||
|
|
||||||
#include <cublas_v2.h>
|
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@@ -19,8 +17,7 @@ class Conv2d {
|
|||||||
int stride,
|
int stride,
|
||||||
std::string padding,
|
std::string padding,
|
||||||
int numFilters,
|
int numFilters,
|
||||||
Activation activation,
|
Activation activation
|
||||||
cublasHandle_t cublasHandle
|
|
||||||
);
|
);
|
||||||
~Conv2d();
|
~Conv2d();
|
||||||
|
|
||||||
@@ -44,7 +41,6 @@ class Conv2d {
|
|||||||
std::vector<float> kernels;
|
std::vector<float> kernels;
|
||||||
|
|
||||||
// Cuda
|
// Cuda
|
||||||
cublasHandle_t cublasHandle;
|
|
||||||
float* d_kernels;
|
float* d_kernels;
|
||||||
float* d_padded;
|
float* d_padded;
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,6 @@
|
|||||||
#ifndef DENSE_LAYER_H
|
#ifndef DENSE_LAYER_H
|
||||||
#define DENSE_LAYER_H
|
#define DENSE_LAYER_H
|
||||||
|
|
||||||
#include <cublas_v2.h>
|
|
||||||
|
|
||||||
#include <functional>
|
#include <functional>
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
@@ -16,8 +14,7 @@ class Dense : public ILayer {
|
|||||||
Dense(
|
Dense(
|
||||||
int inputSize,
|
int inputSize,
|
||||||
int outputSize,
|
int outputSize,
|
||||||
Activation activation,
|
Activation activation
|
||||||
cublasHandle_t cublasHandle
|
|
||||||
);
|
);
|
||||||
~Dense();
|
~Dense();
|
||||||
|
|
||||||
@@ -29,8 +26,6 @@ class Dense : public ILayer {
|
|||||||
int inputSize;
|
int inputSize;
|
||||||
int outputSize;
|
int outputSize;
|
||||||
|
|
||||||
cublasHandle_t cublasHandle;
|
|
||||||
|
|
||||||
float* d_weights;
|
float* d_weights;
|
||||||
float* d_biases;
|
float* d_biases;
|
||||||
|
|
||||||
|
|||||||
@@ -2,8 +2,6 @@
|
|||||||
#ifndef I_LAYER_H
|
#ifndef I_LAYER_H
|
||||||
#define I_LAYER_H
|
#define I_LAYER_H
|
||||||
|
|
||||||
#include <cublas_v2.h>
|
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
namespace Layers {
|
namespace Layers {
|
||||||
|
|||||||
@@ -2,9 +2,6 @@
|
|||||||
#define CUDA_HELPER_H
|
#define CUDA_HELPER_H
|
||||||
|
|
||||||
#include <cuda_runtime.h>
|
#include <cuda_runtime.h>
|
||||||
#include <cublas_v2.h>
|
|
||||||
|
|
||||||
#define IDX2C(i,j,ld) (((j)*(ld))+(i))
|
|
||||||
|
|
||||||
// CUDA error checking macro
|
// CUDA error checking macro
|
||||||
#define CUDA_CHECK(call) \
|
#define CUDA_CHECK(call) \
|
||||||
@@ -18,15 +15,4 @@ do { \
|
|||||||
} \
|
} \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
// cuBLAS error checking macro
|
|
||||||
#define CUBLAS_CHECK(call) \
|
|
||||||
do { \
|
|
||||||
cublasStatus_t result = call; \
|
|
||||||
if (result != CUBLAS_STATUS_SUCCESS) { \
|
|
||||||
fprintf(stderr, "cuBLAS error at %s:%d code=%d\n", \
|
|
||||||
__FILE__, __LINE__, static_cast<unsigned int>(result)); \
|
|
||||||
exit(EXIT_FAILURE); \
|
|
||||||
} \
|
|
||||||
} while (0)
|
|
||||||
|
|
||||||
#endif // CUDA_HELPER_H
|
#endif // CUDA_HELPER_H
|
||||||
|
|||||||
@@ -14,6 +14,8 @@ __global__ void mat_vec_mul_kernel(
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
d_output[tid] = 0.0f;
|
||||||
|
|
||||||
for (int i = 0; i < w; i++) {
|
for (int i = 0; i < w; i++) {
|
||||||
d_output[tid] += d_matrix[tid * w + i] * d_vector[i];
|
d_output[tid] += d_matrix[tid * w + i] * d_vector[i];
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,3 @@
|
|||||||
#include <cublas_v2.h>
|
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
|
|
||||||
#include "activations.cuh"
|
#include "activations.cuh"
|
||||||
@@ -14,15 +12,13 @@ Layers::Conv2d::Conv2d(
|
|||||||
int stride,
|
int stride,
|
||||||
std::string padding,
|
std::string padding,
|
||||||
int numFilters,
|
int numFilters,
|
||||||
Activation activation,
|
Activation activation
|
||||||
cublasHandle_t cublasHandle
|
|
||||||
)
|
)
|
||||||
: inputSize(inputSize),
|
: inputSize(inputSize),
|
||||||
inputChannels(inputChannels),
|
inputChannels(inputChannels),
|
||||||
kernelSize(kernelSize),
|
kernelSize(kernelSize),
|
||||||
stride(stride),
|
stride(stride),
|
||||||
numFilters(numFilters),
|
numFilters(numFilters),
|
||||||
cublasHandle(cublasHandle),
|
|
||||||
activation(activation) {
|
activation(activation) {
|
||||||
// Allocate memory for kernels
|
// Allocate memory for kernels
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
#include <cublas_v2.h>
|
|
||||||
#include <cuda_runtime.h>
|
#include <cuda_runtime.h>
|
||||||
|
|
||||||
#include <cstdio>
|
#include <cstdio>
|
||||||
@@ -9,16 +8,15 @@
|
|||||||
#include "activations.cuh"
|
#include "activations.cuh"
|
||||||
#include "cuda_helper.cuh"
|
#include "cuda_helper.cuh"
|
||||||
#include "dense.cuh"
|
#include "dense.cuh"
|
||||||
|
#include "matrix_math.cuh"
|
||||||
|
|
||||||
Layers::Dense::Dense(
|
Layers::Dense::Dense(
|
||||||
int inputSize,
|
int inputSize,
|
||||||
int outputSize,
|
int outputSize,
|
||||||
Activation activation,
|
Activation activation
|
||||||
cublasHandle_t cublasHandle
|
|
||||||
)
|
)
|
||||||
: inputSize(inputSize),
|
: inputSize(inputSize),
|
||||||
outputSize(outputSize),
|
outputSize(outputSize),
|
||||||
cublasHandle(cublasHandle),
|
|
||||||
activation(activation) {
|
activation(activation) {
|
||||||
// Allocate memory for weights and biases
|
// Allocate memory for weights and biases
|
||||||
weights.resize(outputSize * inputSize);
|
weights.resize(outputSize * inputSize);
|
||||||
@@ -54,35 +52,30 @@ void Layers::Dense::initializeBiases() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void Layers::Dense::forward(const float* d_input, float* d_output) {
|
void Layers::Dense::forward(const float* d_input, float* d_output) {
|
||||||
const float alpha = 1.0f;
|
|
||||||
const float beta = 0.0f;
|
|
||||||
|
|
||||||
CUBLAS_CHECK(cublasSgemv(
|
mat_vec_mul_kernel<<<1, outputSize>>>(
|
||||||
cublasHandle, CUBLAS_OP_N, outputSize, inputSize, &alpha, d_weights,
|
d_weights, d_input, d_output, inputSize, outputSize
|
||||||
outputSize, d_input, 1, &beta, d_output, 1
|
|
||||||
));
|
|
||||||
CUBLAS_CHECK(
|
|
||||||
cublasSaxpy(cublasHandle, outputSize, &alpha, d_biases, 1, d_output, 1)
|
|
||||||
);
|
);
|
||||||
|
|
||||||
int threadsPerBlock = 256;
|
vec_vec_add_kernel<<<1, outputSize>>>(
|
||||||
int blocksPerGrid = (outputSize + threadsPerBlock - 1) / threadsPerBlock;
|
d_biases, d_output, d_output, outputSize
|
||||||
|
);
|
||||||
|
|
||||||
switch (activation) {
|
switch (activation) {
|
||||||
case SIGMOID:
|
case SIGMOID:
|
||||||
sigmoid_kernel<<<blocksPerGrid, threadsPerBlock>>>(
|
sigmoid_kernel<<<1, outputSize>>>(
|
||||||
d_output, d_output, outputSize
|
d_output, d_output, outputSize
|
||||||
);
|
);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case RELU:
|
case RELU:
|
||||||
relu_kernel<<<blocksPerGrid, threadsPerBlock>>>(
|
relu_kernel<<<1, outputSize>>>(
|
||||||
d_output, d_output, outputSize
|
d_output, d_output, outputSize
|
||||||
);
|
);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
linear_kernel<<<blocksPerGrid, threadsPerBlock>>>(
|
linear_kernel<<<1, outputSize>>>(
|
||||||
d_output, d_output, outputSize
|
d_output, d_output, outputSize
|
||||||
);
|
);
|
||||||
break;
|
break;
|
||||||
@@ -92,12 +85,13 @@ void Layers::Dense::forward(const float* d_input, float* d_output) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void Layers::Dense::toCuda() {
|
void Layers::Dense::toCuda() {
|
||||||
CUBLAS_CHECK(cublasSetMatrix(
|
CUDA_CHECK(cudaMemcpy(
|
||||||
outputSize, inputSize, sizeof(float), weights.data(), outputSize,
|
d_weights, weights.data(), sizeof(float) * inputSize * outputSize,
|
||||||
d_weights, outputSize
|
cudaMemcpyHostToDevice
|
||||||
));
|
));
|
||||||
CUBLAS_CHECK(cublasSetVector(
|
CUDA_CHECK(cudaMemcpy(
|
||||||
biases.size(), sizeof(float), biases.data(), 1, d_biases, 1
|
d_biases, biases.data(), sizeof(float) * outputSize,
|
||||||
|
cudaMemcpyHostToDevice
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -111,10 +105,9 @@ void Layers::Dense::setWeights(
|
|||||||
exit(EXIT_FAILURE);
|
exit(EXIT_FAILURE);
|
||||||
}
|
}
|
||||||
|
|
||||||
for (int j = 0; j < inputSize; ++j) {
|
|
||||||
for (int i = 0; i < outputSize; ++i) {
|
for (int i = 0; i < outputSize; ++i) {
|
||||||
int idx = IDX2C(i, j, outputSize);
|
for (int j = 0; j < inputSize; ++j) {
|
||||||
weights[idx] = weights_input[i][j];
|
weights[i * inputSize + j] = weights_input[i][j];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
#include <cublas_v2.h>
|
|
||||||
#include <cuda_runtime.h>
|
#include <cuda_runtime.h>
|
||||||
|
|
||||||
#include <cstdio>
|
#include <cstdio>
|
||||||
@@ -6,7 +5,7 @@
|
|||||||
|
|
||||||
#include "cuda_helper.cuh"
|
#include "cuda_helper.cuh"
|
||||||
|
|
||||||
cudaDeviceProp initializeCUDA(cublasHandle_t& cublasHandle) {
|
cudaDeviceProp initializeCUDA() {
|
||||||
int deviceCount;
|
int deviceCount;
|
||||||
CUDA_CHECK(cudaGetDeviceCount(&deviceCount));
|
CUDA_CHECK(cudaGetDeviceCount(&deviceCount));
|
||||||
|
|
||||||
@@ -23,8 +22,5 @@ cudaDeviceProp initializeCUDA(cublasHandle_t& cublasHandle) {
|
|||||||
|
|
||||||
std::printf("Using CUDA device %d: %s\n", device, deviceProp.name);
|
std::printf("Using CUDA device %d: %s\n", device, deviceProp.name);
|
||||||
|
|
||||||
// Initialize cuBLAS
|
|
||||||
CUBLAS_CHECK(cublasCreate(&cublasHandle));
|
|
||||||
|
|
||||||
return deviceProp;
|
return deviceProp;
|
||||||
}
|
}
|
||||||
@@ -7,12 +7,6 @@ add_executable(test_main
|
|||||||
kernels/test_padding.cu
|
kernels/test_padding.cu
|
||||||
)
|
)
|
||||||
|
|
||||||
add_library(test_utils
|
target_link_libraries(test_main ${GTEST_BOTH_LIBRARIES} CUDANet)
|
||||||
test_utils/test_cublas_fixture.cu
|
|
||||||
)
|
|
||||||
|
|
||||||
target_include_directories(test_utils PUBLIC test_utils)
|
|
||||||
|
|
||||||
target_link_libraries(test_main ${GTEST_BOTH_LIBRARIES} CUDANet test_utils)
|
|
||||||
|
|
||||||
add_test(NAME TestMain COMMAND test_main)
|
add_test(NAME TestMain COMMAND test_main)
|
||||||
@@ -4,15 +4,11 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
#include "activations.cuh"
|
#include "activations.cuh"
|
||||||
#include "test_cublas_fixture.cuh"
|
|
||||||
|
|
||||||
class ActivationsTest : public CublasTestFixture {
|
TEST(ActivationsTest, SigmoidSanityCheck) {
|
||||||
protected:
|
|
||||||
cudaError_t cudaStatus;
|
cudaError_t cudaStatus;
|
||||||
cublasStatus_t cublasStatus;
|
|
||||||
};
|
|
||||||
|
|
||||||
TEST_F(ActivationsTest, SigmoidSanityCheck) {
|
|
||||||
float input[3] = {-100.0f, 0.0f, 100.0f};
|
float input[3] = {-100.0f, 0.0f, 100.0f};
|
||||||
|
|
||||||
std::vector<float> expected_output = {0.0f, 0.5f, 1.0f};
|
std::vector<float> expected_output = {0.0f, 0.5f, 1.0f};
|
||||||
@@ -26,8 +22,8 @@ TEST_F(ActivationsTest, SigmoidSanityCheck) {
|
|||||||
cudaStatus = cudaMalloc((void**)&d_output, sizeof(float) * 3);
|
cudaStatus = cudaMalloc((void**)&d_output, sizeof(float) * 3);
|
||||||
EXPECT_EQ(cudaStatus, cudaSuccess);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
cublasStatus = cublasSetVector(3, sizeof(float), input, 1, d_input, 1);
|
cudaStatus = cudaMemcpy(d_input, input, sizeof(float) * 3, cudaMemcpyHostToDevice);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
sigmoid_kernel<<<1, 3>>>(d_input, d_output, 3);
|
sigmoid_kernel<<<1, 3>>>(d_input, d_output, 3);
|
||||||
cudaStatus = cudaDeviceSynchronize();
|
cudaStatus = cudaDeviceSynchronize();
|
||||||
@@ -35,9 +31,8 @@ TEST_F(ActivationsTest, SigmoidSanityCheck) {
|
|||||||
|
|
||||||
std::vector<float> output(3);
|
std::vector<float> output(3);
|
||||||
|
|
||||||
cublasStatus =
|
cudaStatus = cudaMemcpy(output.data(), d_output, sizeof(float) * 3, cudaMemcpyDeviceToHost);
|
||||||
cublasGetVector(3, sizeof(float), d_output, 1, output.data(), 1);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
|
||||||
|
|
||||||
for (int i = 0; i < 3; i++) {
|
for (int i = 0; i < 3; i++) {
|
||||||
EXPECT_NEAR(expected_output[i], output[i], 1e-5);
|
EXPECT_NEAR(expected_output[i], output[i], 1e-5);
|
||||||
|
|||||||
@@ -4,15 +4,10 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
#include "padding.cuh"
|
#include "padding.cuh"
|
||||||
#include "test_cublas_fixture.cuh"
|
|
||||||
|
|
||||||
class PaddingTest : public CublasTestFixture {
|
TEST(PaddingTest, SimplePaddingTest) {
|
||||||
protected:
|
|
||||||
cudaError_t cudaStatus;
|
cudaError_t cudaStatus;
|
||||||
cublasStatus_t cublasStatus;
|
|
||||||
};
|
|
||||||
|
|
||||||
TEST_F(PaddingTest, SimplePaddingTest) {
|
|
||||||
int w = 2;
|
int w = 2;
|
||||||
int h = 3;
|
int h = 3;
|
||||||
int n = 2;
|
int n = 2;
|
||||||
@@ -48,9 +43,10 @@ TEST_F(PaddingTest, SimplePaddingTest) {
|
|||||||
std::vector<float> input = {0.0f, 2.0f, 4.0f, 1.0f, 3.0f, 5.0f,
|
std::vector<float> input = {0.0f, 2.0f, 4.0f, 1.0f, 3.0f, 5.0f,
|
||||||
6.0f, 8.0f, 10.0f, 7.0f, 9.0f, 11.0f};
|
6.0f, 8.0f, 10.0f, 7.0f, 9.0f, 11.0f};
|
||||||
|
|
||||||
cublasStatus =
|
cudaStatus = cudaMemcpy(
|
||||||
cublasSetVector(inputSize, sizeof(float), input.data(), 1, d_input, 1);
|
d_input, input.data(), sizeof(float) * inputSize, cudaMemcpyHostToDevice
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
int THREADS_PER_BLOCK = 64;
|
int THREADS_PER_BLOCK = 64;
|
||||||
int BLOCKS = paddedSize / THREADS_PER_BLOCK + 1;
|
int BLOCKS = paddedSize / THREADS_PER_BLOCK + 1;
|
||||||
@@ -69,9 +65,12 @@ TEST_F(PaddingTest, SimplePaddingTest) {
|
|||||||
};
|
};
|
||||||
|
|
||||||
std::vector<float> output(paddedSize);
|
std::vector<float> output(paddedSize);
|
||||||
cublasStatus = cublasGetVector(
|
|
||||||
paddedSize, sizeof(float), d_padded, 1, output.data(), 1
|
cudaStatus = cudaMemcpy(
|
||||||
|
output.data(), d_padded, sizeof(float) * paddedSize,
|
||||||
|
cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
for (int i = 0; i < paddedSize; i++) {
|
for (int i = 0; i < paddedSize; i++) {
|
||||||
EXPECT_NEAR(expectedOutput[i], output[i], 1e-5);
|
EXPECT_NEAR(expectedOutput[i], output[i], 1e-5);
|
||||||
|
|||||||
@@ -5,9 +5,9 @@
|
|||||||
|
|
||||||
#include "activations.cuh"
|
#include "activations.cuh"
|
||||||
#include "dense.cuh"
|
#include "dense.cuh"
|
||||||
#include "test_cublas_fixture.cuh"
|
|
||||||
|
|
||||||
class DenseLayerTest : public CublasTestFixture {
|
|
||||||
|
class DenseLayerTest : public::testing::Test {
|
||||||
protected:
|
protected:
|
||||||
Layers::Dense commonTestSetup(
|
Layers::Dense commonTestSetup(
|
||||||
int inputSize,
|
int inputSize,
|
||||||
@@ -21,7 +21,7 @@ class DenseLayerTest : public CublasTestFixture {
|
|||||||
) {
|
) {
|
||||||
// Create Dense layer
|
// Create Dense layer
|
||||||
Layers::Dense denseLayer(
|
Layers::Dense denseLayer(
|
||||||
inputSize, outputSize, activation, cublasHandle
|
inputSize, outputSize, activation
|
||||||
);
|
);
|
||||||
|
|
||||||
// Set weights and biases
|
// Set weights and biases
|
||||||
@@ -36,10 +36,11 @@ class DenseLayerTest : public CublasTestFixture {
|
|||||||
EXPECT_EQ(cudaStatus, cudaSuccess);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
// Copy input to device
|
// Copy input to device
|
||||||
cublasStatus = cublasSetVector(
|
cudaStatus = cudaMemcpy(
|
||||||
input.size(), sizeof(float), input.data(), 1, d_input, 1
|
d_input, input.data(), sizeof(float) * input.size(), cudaMemcpyHostToDevice
|
||||||
);
|
);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
|
|
||||||
return denseLayer;
|
return denseLayer;
|
||||||
}
|
}
|
||||||
@@ -51,7 +52,6 @@ class DenseLayerTest : public CublasTestFixture {
|
|||||||
}
|
}
|
||||||
|
|
||||||
cudaError_t cudaStatus;
|
cudaError_t cudaStatus;
|
||||||
cublasStatus_t cublasStatus;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
TEST_F(DenseLayerTest, Init) {
|
TEST_F(DenseLayerTest, Init) {
|
||||||
@@ -60,10 +60,8 @@ TEST_F(DenseLayerTest, Init) {
|
|||||||
int inputSize = i;
|
int inputSize = i;
|
||||||
int outputSize = j;
|
int outputSize = j;
|
||||||
|
|
||||||
// std::cout << "Dense layer: input size = " << inputSize << ",
|
|
||||||
// output size = " << outputSize << std::endl;
|
|
||||||
Layers::Dense denseLayer(
|
Layers::Dense denseLayer(
|
||||||
inputSize, outputSize, SIGMOID, cublasHandle
|
inputSize, outputSize, SIGMOID
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -81,7 +79,7 @@ TEST_F(DenseLayerTest, setWeights) {
|
|||||||
{1.3f, 0.5f, 0.0f, 1.7f}
|
{1.3f, 0.5f, 0.0f, 1.7f}
|
||||||
};
|
};
|
||||||
|
|
||||||
Layers::Dense denseLayer(inputSize, outputSize, SIGMOID, cublasHandle);
|
Layers::Dense denseLayer(inputSize, outputSize, SIGMOID);
|
||||||
|
|
||||||
denseLayer.setWeights(weights);
|
denseLayer.setWeights(weights);
|
||||||
}
|
}
|
||||||
@@ -113,10 +111,10 @@ TEST_F(DenseLayerTest, ForwardUnitWeightMatrixLinear) {
|
|||||||
denseLayer.forward(d_input, d_output);
|
denseLayer.forward(d_input, d_output);
|
||||||
|
|
||||||
std::vector<float> output(outputSize);
|
std::vector<float> output(outputSize);
|
||||||
cublasStatus = cublasGetVector(
|
cudaStatus = cudaMemcpy(
|
||||||
outputSize, sizeof(float), d_output, 1, output.data(), 1
|
output.data(), d_output, sizeof(float) * outputSize, cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
// Check if the output is a zero vector
|
// Check if the output is a zero vector
|
||||||
EXPECT_FLOAT_EQ(output[0], 2.0f);
|
EXPECT_FLOAT_EQ(output[0], 2.0f);
|
||||||
@@ -150,10 +148,10 @@ TEST_F(DenseLayerTest, ForwardRandomWeightMatrixRelu) {
|
|||||||
denseLayer.forward(d_input, d_output);
|
denseLayer.forward(d_input, d_output);
|
||||||
|
|
||||||
std::vector<float> output(outputSize);
|
std::vector<float> output(outputSize);
|
||||||
cublasStatus = cublasGetVector(
|
cudaStatus = cudaMemcpy(
|
||||||
outputSize, sizeof(float), d_output, 1, output.data(), 1
|
output.data(), d_output, sizeof(float) * outputSize, cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
// weights * inputs = 0.1, 12.5, 8.3, -2.2
|
// weights * inputs = 0.1, 12.5, 8.3, -2.2
|
||||||
// + biases = 0.3, 13, 9, -3.3
|
// + biases = 0.3, 13, 9, -3.3
|
||||||
@@ -193,10 +191,10 @@ TEST_F(DenseLayerTest, ForwardRandomWeightMatrixSigmoid) {
|
|||||||
denseLayer.forward(d_input, d_output);
|
denseLayer.forward(d_input, d_output);
|
||||||
|
|
||||||
std::vector<float> output(outputSize);
|
std::vector<float> output(outputSize);
|
||||||
cublasStatus = cublasGetVector(
|
cudaStatus = cudaMemcpy(
|
||||||
outputSize, sizeof(float), d_output, 1, output.data(), 1
|
output.data(), d_output, sizeof(float) * outputSize, cudaMemcpyDeviceToHost
|
||||||
);
|
);
|
||||||
EXPECT_EQ(cublasStatus, CUBLAS_STATUS_SUCCESS);
|
EXPECT_EQ(cudaStatus, cudaSuccess);
|
||||||
|
|
||||||
// weights * input = 0.95, 0.43, 0.45, 0.93
|
// weights * input = 0.95, 0.43, 0.45, 0.93
|
||||||
// + biases = 1.05, 0.63, 0.75, 1.33
|
// + biases = 1.05, 0.63, 0.75, 1.33
|
||||||
|
|||||||
@@ -1,14 +0,0 @@
|
|||||||
#include <cublas_v2.h>
|
|
||||||
#include <gtest/gtest.h>
|
|
||||||
|
|
||||||
#include "test_cublas_fixture.cuh"
|
|
||||||
|
|
||||||
cublasHandle_t CublasTestFixture::cublasHandle;
|
|
||||||
|
|
||||||
void CublasTestFixture::SetUpTestSuite() {
|
|
||||||
cublasCreate(&cublasHandle);
|
|
||||||
}
|
|
||||||
|
|
||||||
void CublasTestFixture::TearDownTestSuite() {
|
|
||||||
cublasDestroy(cublasHandle);
|
|
||||||
}
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
#include <cublas_v2.h>
|
|
||||||
#include <gtest/gtest.h>
|
|
||||||
|
|
||||||
class CublasTestFixture : public ::testing::Test {
|
|
||||||
protected:
|
|
||||||
static cublasHandle_t cublasHandle;
|
|
||||||
|
|
||||||
static void SetUpTestSuite();
|
|
||||||
static void TearDownTestSuite();
|
|
||||||
};
|
|
||||||
Reference in New Issue
Block a user