Loading neuralnetworks/1.2/vts/functional/Android.bp +12 −1 Original line number Diff line number Diff line Loading @@ -20,7 +20,11 @@ cc_test { defaults: ["VtsHalNeuralNetworksTargetTestDefaults"], srcs: [ "GeneratedTestsV1_0.cpp", ] ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], } // Tests for V1_1 models using the V1_2 HAL. Loading @@ -30,6 +34,10 @@ cc_test { srcs: [ "GeneratedTestsV1_1.cpp", ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], } // Tests for V1_2 models. Loading @@ -40,5 +48,8 @@ cc_test { "BasicTests.cpp", "GeneratedTests.cpp", ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], } neuralnetworks/1.2/vts/functional/GeneratedTests.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading neuralnetworks/1.2/vts/functional/GeneratedTestsV1_0.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading neuralnetworks/1.2/vts/functional/GeneratedTestsV1_1.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading neuralnetworks/1.2/vts/functional/ValidateModel.cpp +3 −0 Original line number Diff line number Diff line Loading @@ -172,6 +172,9 @@ static uint32_t getInvalidRank(OperandType type) { static void mutateOperandRankTest(const sp<IDevice>& device, const Model& model) { for (size_t operand = 0; operand < model.operands.size(); ++operand) { const uint32_t invalidRank = getInvalidRank(model.operands[operand].type); if (invalidRank == 0) { continue; } const std::string message = "mutateOperandRankTest: operand " + std::to_string(operand) + " has rank of " + std::to_string(invalidRank); validate(device, message, model, [operand, invalidRank](Model* model) { Loading Loading
neuralnetworks/1.2/vts/functional/Android.bp +12 −1 Original line number Diff line number Diff line Loading @@ -20,7 +20,11 @@ cc_test { defaults: ["VtsHalNeuralNetworksTargetTestDefaults"], srcs: [ "GeneratedTestsV1_0.cpp", ] ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], } // Tests for V1_1 models using the V1_2 HAL. Loading @@ -30,6 +34,10 @@ cc_test { srcs: [ "GeneratedTestsV1_1.cpp", ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], } // Tests for V1_2 models. Loading @@ -40,5 +48,8 @@ cc_test { "BasicTests.cpp", "GeneratedTests.cpp", ], cflags: [ "-DNN_TEST_DYNAMIC_OUTPUT_SHAPE" ], test_suites: ["general-tests"], }
neuralnetworks/1.2/vts/functional/GeneratedTests.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading
neuralnetworks/1.2/vts/functional/GeneratedTestsV1_0.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading
neuralnetworks/1.2/vts/functional/GeneratedTestsV1_1.cpp +2 −1 Original line number Diff line number Diff line Loading @@ -33,7 +33,8 @@ namespace neuralnetworks { namespace generated_tests { using ::test_helper::MixedTypedExample; extern void Execute(const sp<V1_2::IDevice>&, std::function<V1_2::Model(void)>, std::function<bool(int)>, const std::vector<MixedTypedExample>&); std::function<bool(int)>, const std::vector<MixedTypedExample>&, bool testDynamicOutputShape = false); } // namespace generated_tests namespace V1_2 { Loading
neuralnetworks/1.2/vts/functional/ValidateModel.cpp +3 −0 Original line number Diff line number Diff line Loading @@ -172,6 +172,9 @@ static uint32_t getInvalidRank(OperandType type) { static void mutateOperandRankTest(const sp<IDevice>& device, const Model& model) { for (size_t operand = 0; operand < model.operands.size(); ++operand) { const uint32_t invalidRank = getInvalidRank(model.operands[operand].type); if (invalidRank == 0) { continue; } const std::string message = "mutateOperandRankTest: operand " + std::to_string(operand) + " has rank of " + std::to_string(invalidRank); validate(device, message, model, [operand, invalidRank](Model* model) { Loading