Spaces:
Runtime error
Runtime error
// SPDX-License-Identifier: Apache-2.0 | |
TEST(TestPushConstants, TestConstantsAlgoDispatchOverride) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
float x; | |
float y; | |
float z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { float pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<float>> tensor = | |
mgr.tensor({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = mgr.algorithm( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { 0.0, 0.0, 0.0 }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>(algo, | |
std::vector<float>{ 0.1, 0.2, 0.3 }); | |
sq->eval<kp::OpAlgoDispatch>(algo, | |
std::vector<float>{ 0.3, 0.2, 0.1 }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ(tensor->vector(), std::vector<float>({ 0.4, 0.4, 0.4 })); | |
} | |
} | |
} | |
TEST(TestPushConstants, TestConstantsAlgoDispatchNoOverride) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
float x; | |
float y; | |
float z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { float pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<float>> tensor = | |
mgr.tensor({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = mgr.algorithm( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { 0.1, 0.2, 0.3 }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>(algo); | |
sq->eval<kp::OpAlgoDispatch>(algo, | |
std::vector<float>{ 0.3, 0.2, 0.1 }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ(tensor->vector(), std::vector<float>({ 0.4, 0.4, 0.4 })); | |
} | |
} | |
} | |
TEST(TestPushConstants, TestConstantsWrongSize) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
float x; | |
float y; | |
float z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { float pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<float>> tensor = | |
mgr.tensor({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = mgr.algorithm( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { 0.0 }); | |
sq = mgr.sequence()->record<kp::OpTensorSyncDevice>({ tensor }); | |
EXPECT_THROW(sq->record<kp::OpAlgoDispatch>( | |
algo, std::vector<float>{ 0.1, 0.2, 0.3 }), | |
std::runtime_error); | |
} | |
} | |
} | |
// TODO: Ensure different types are considered for push constants | |
// TEST(TestPushConstants, TestConstantsWrongType) | |
// { | |
// { | |
// std::string shader(R"( | |
// #version 450 | |
// layout(push_constant) uniform PushConstants { | |
// float x; | |
// float y; | |
// float z; | |
// } pcs; | |
// layout (local_size_x = 1) in; | |
// layout(set = 0, binding = 0) buffer a { float pa[]; }; | |
// void main() { | |
// pa[0] += pcs.x; | |
// pa[1] += pcs.y; | |
// pa[2] += pcs.z; | |
// })"); | |
// | |
// std::vector<uint32_t> spirv = compileSource(shader); | |
// | |
// std::shared_ptr<kp::Sequence> sq = nullptr; | |
// | |
// { | |
// kp::Manager mgr; | |
// | |
// std::shared_ptr<kp::TensorT<float>> tensor = | |
// mgr.tensor({ 0, 0, 0 }); | |
// | |
// std::shared_ptr<kp::Algorithm> algo = mgr.algorithm( | |
// { tensor }, spirv, kp::Workgroup({ 1 }), {}, { 0.0 }); | |
// | |
// sq = mgr.sequence()->record<kp::OpTensorSyncDevice>({ tensor }); | |
// | |
// EXPECT_THROW(sq->record<kp::OpAlgoDispatch>( | |
// algo, std::vector<uint32_t>{ 1, 2, 3 }), | |
// std::runtime_error); | |
// } | |
// } | |
// } | |
TEST(TestPushConstants, TestConstantsMixedTypes) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
float x; | |
uint y; | |
int z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { float pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y - 2147483000; | |
pa[2] += pcs.z; | |
})"); | |
struct TestConsts | |
{ | |
float x; | |
uint32_t y; | |
int32_t z; | |
}; | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<float>> tensor = | |
mgr.tensorT<float>({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = | |
mgr.algorithm<float, TestConsts>( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { { 0, 0, 0 } }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, std::vector<TestConsts>{ { 15.32, 2147483650, 10 } }); | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, std::vector<TestConsts>{ { 30.32, 2147483650, -3 } }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ(tensor->vector(), std::vector<float>({ 45.64, 1300, 7 })); | |
} | |
} | |
} | |
TEST(TestPushConstants, TestConstantsInt) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
int x; | |
int y; | |
int z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { int pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<int32_t>> tensor = | |
mgr.tensorT<int32_t>({ -1, -1, -1 }); | |
std::shared_ptr<kp::Algorithm> algo = | |
mgr.algorithm<int32_t, int32_t>( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { { 0, 0, 0 } }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, std::vector<int32_t>{ { -1, -1, -1 } }); | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, std::vector<int32_t>{ { -1, -1, -1 } }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ(tensor->vector(), std::vector<int32_t>({ -3, -3, -3 })); | |
} | |
} | |
} | |
TEST(TestPushConstants, TestConstantsUnsignedInt) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
uint x; | |
uint y; | |
uint z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { uint pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<uint32_t>> tensor = | |
mgr.tensorT<uint32_t>({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = | |
mgr.algorithm<uint32_t, uint32_t>( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { { 0, 0, 0 } }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, | |
std::vector<uint32_t>{ { 2147483650, 2147483650, 2147483650 } }); | |
sq->eval<kp::OpAlgoDispatch>(algo, | |
std::vector<uint32_t>{ { 5, 5, 5 } }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ( | |
tensor->vector(), | |
std::vector<uint32_t>({ 2147483655, 2147483655, 2147483655 })); | |
} | |
} | |
} | |
TEST(TestPushConstants, TestConstantsDouble) | |
{ | |
{ | |
std::string shader(R"( | |
#version 450 | |
layout(push_constant) uniform PushConstants { | |
double x; | |
double y; | |
double z; | |
} pcs; | |
layout (local_size_x = 1) in; | |
layout(set = 0, binding = 0) buffer a { double pa[]; }; | |
void main() { | |
pa[0] += pcs.x; | |
pa[1] += pcs.y; | |
pa[2] += pcs.z; | |
})"); | |
std::vector<uint32_t> spirv = compileSource(shader); | |
std::shared_ptr<kp::Sequence> sq = nullptr; | |
{ | |
kp::Manager mgr; | |
std::shared_ptr<kp::TensorT<double>> tensor = | |
mgr.tensorT<double>({ 0, 0, 0 }); | |
std::shared_ptr<kp::Algorithm> algo = mgr.algorithm<double, double>( | |
{ tensor }, spirv, kp::Workgroup({ 1 }), {}, { { 0, 0, 0 } }); | |
sq = mgr.sequence()->eval<kp::OpTensorSyncDevice>({ tensor }); | |
// We need to run this in sequence to avoid race condition | |
// We can't use atomicAdd as swiftshader doesn't support it for | |
// float | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, | |
std::vector<double>{ { 1.1111222233334444, | |
2.1111222233334444, | |
3.1111222233334444 } }); | |
sq->eval<kp::OpAlgoDispatch>( | |
algo, | |
std::vector<double>{ { 1.1111222233334444, | |
2.1111222233334444, | |
3.1111222233334444 } }); | |
sq->eval<kp::OpTensorSyncLocal>({ tensor }); | |
EXPECT_EQ(tensor->vector(), | |
std::vector<double>({ 2.2222444466668888, | |
4.2222444466668888, | |
6.2222444466668888 })); | |
} | |
} | |
} | |