Loading neuralnetworks/1.2/vts/functional/GeneratedTestHarness.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -272,7 +272,7 @@ void EvaluatePreparedModel(const sp<IPreparedModel>& preparedModel, const TestMo int n; std::tie(n, outputShapes, timing, std::ignore) = controller->compute(request, testConfig.measureTiming, keys); executionStatus = nn::convertResultCodeToErrorStatus(n); executionStatus = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(n)); break; } Loading neuralnetworks/1.2/vts/functional/ValidateBurst.cpp +3 −2 Original line number Diff line number Diff line Loading @@ -296,7 +296,8 @@ static void validateBurstFmqLength(const sp<IPreparedModel>& preparedModel, // collect serialized result by running regular burst const auto [nRegular, outputShapesRegular, timingRegular, fallbackRegular] = controllerRegular->compute(request, MeasureTiming::NO, keys); const ErrorStatus statusRegular = nn::convertResultCodeToErrorStatus(nRegular); const ErrorStatus statusRegular = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(nRegular)); EXPECT_FALSE(fallbackRegular); // skip test if regular burst output isn't useful for testing a failure Loading @@ -312,7 +313,7 @@ static void validateBurstFmqLength(const sp<IPreparedModel>& preparedModel, // large enough to return the serialized result const auto [nSmall, outputShapesSmall, timingSmall, fallbackSmall] = controllerSmall->compute(request, MeasureTiming::NO, keys); const ErrorStatus statusSmall = nn::convertResultCodeToErrorStatus(nSmall); const ErrorStatus statusSmall = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(nSmall)); EXPECT_NE(ErrorStatus::NONE, statusSmall); EXPECT_EQ(0u, outputShapesSmall.size()); EXPECT_TRUE(badTiming(timingSmall)); Loading neuralnetworks/1.2/vts/functional/ValidateRequest.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -107,7 +107,7 @@ static void validate(const sp<IPreparedModel>& preparedModel, const std::string& // execute and verify const auto [n, outputShapes, timing, fallback] = burst->compute(request, measure, keys); const ErrorStatus status = nn::convertResultCodeToErrorStatus(n); const ErrorStatus status = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(n)); EXPECT_EQ(ErrorStatus::INVALID_ARGUMENT, status); EXPECT_EQ(outputShapes.size(), 0); EXPECT_TRUE(badTiming(timing)); Loading neuralnetworks/1.3/vts/functional/Android.bp +6 −5 Original line number Diff line number Diff line Loading @@ -15,11 +15,12 @@ // cc_library_static { name: "VtsHalNeuralNetworksV1_3Callbacks", name: "VtsHalNeuralNetworksV1_3_utils", defaults: ["VtsHalTargetTestDefaults"], export_include_dirs: ["include"], srcs: [ "Callbacks.cpp", "Utils.cpp", ], static_libs: [ "android.hardware.neuralnetworks@1.0", Loading @@ -29,7 +30,7 @@ cc_library_static { ], header_libs: [ "libbase_headers", ] ], } cc_test { Loading @@ -50,6 +51,9 @@ cc_test { "libnativewindow", ], static_libs: [ "VtsHalNeuralNetworksV1_0_utils", "VtsHalNeuralNetworksV1_2Callbacks", "VtsHalNeuralNetworksV1_3_utils", "android.hardware.neuralnetworks@1.0", "android.hardware.neuralnetworks@1.1", "android.hardware.neuralnetworks@1.2", Loading @@ -60,9 +64,6 @@ cc_test { "libhidlmemory", "libneuralnetworks_generated_test_harness", "libneuralnetworks_utils", "VtsHalNeuralNetworksV1_0_utils", "VtsHalNeuralNetworksV1_2Callbacks", "VtsHalNeuralNetworksV1_3Callbacks", ], whole_static_libs: [ "neuralnetworks_generated_V1_0_example", Loading neuralnetworks/1.3/vts/functional/BasicTests.cpp +0 −1 Original line number Diff line number Diff line Loading @@ -21,7 +21,6 @@ namespace android::hardware::neuralnetworks::V1_3::vts::functional { using V1_0::DeviceStatus; using V1_0::ErrorStatus; using V1_0::PerformanceInfo; using V1_2::Constant; using V1_2::DeviceType; Loading Loading
neuralnetworks/1.2/vts/functional/GeneratedTestHarness.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -272,7 +272,7 @@ void EvaluatePreparedModel(const sp<IPreparedModel>& preparedModel, const TestMo int n; std::tie(n, outputShapes, timing, std::ignore) = controller->compute(request, testConfig.measureTiming, keys); executionStatus = nn::convertResultCodeToErrorStatus(n); executionStatus = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(n)); break; } Loading
neuralnetworks/1.2/vts/functional/ValidateBurst.cpp +3 −2 Original line number Diff line number Diff line Loading @@ -296,7 +296,8 @@ static void validateBurstFmqLength(const sp<IPreparedModel>& preparedModel, // collect serialized result by running regular burst const auto [nRegular, outputShapesRegular, timingRegular, fallbackRegular] = controllerRegular->compute(request, MeasureTiming::NO, keys); const ErrorStatus statusRegular = nn::convertResultCodeToErrorStatus(nRegular); const ErrorStatus statusRegular = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(nRegular)); EXPECT_FALSE(fallbackRegular); // skip test if regular burst output isn't useful for testing a failure Loading @@ -312,7 +313,7 @@ static void validateBurstFmqLength(const sp<IPreparedModel>& preparedModel, // large enough to return the serialized result const auto [nSmall, outputShapesSmall, timingSmall, fallbackSmall] = controllerSmall->compute(request, MeasureTiming::NO, keys); const ErrorStatus statusSmall = nn::convertResultCodeToErrorStatus(nSmall); const ErrorStatus statusSmall = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(nSmall)); EXPECT_NE(ErrorStatus::NONE, statusSmall); EXPECT_EQ(0u, outputShapesSmall.size()); EXPECT_TRUE(badTiming(timingSmall)); Loading
neuralnetworks/1.2/vts/functional/ValidateRequest.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -107,7 +107,7 @@ static void validate(const sp<IPreparedModel>& preparedModel, const std::string& // execute and verify const auto [n, outputShapes, timing, fallback] = burst->compute(request, measure, keys); const ErrorStatus status = nn::convertResultCodeToErrorStatus(n); const ErrorStatus status = nn::convertToV1_0(nn::convertResultCodeToErrorStatus(n)); EXPECT_EQ(ErrorStatus::INVALID_ARGUMENT, status); EXPECT_EQ(outputShapes.size(), 0); EXPECT_TRUE(badTiming(timing)); Loading
neuralnetworks/1.3/vts/functional/Android.bp +6 −5 Original line number Diff line number Diff line Loading @@ -15,11 +15,12 @@ // cc_library_static { name: "VtsHalNeuralNetworksV1_3Callbacks", name: "VtsHalNeuralNetworksV1_3_utils", defaults: ["VtsHalTargetTestDefaults"], export_include_dirs: ["include"], srcs: [ "Callbacks.cpp", "Utils.cpp", ], static_libs: [ "android.hardware.neuralnetworks@1.0", Loading @@ -29,7 +30,7 @@ cc_library_static { ], header_libs: [ "libbase_headers", ] ], } cc_test { Loading @@ -50,6 +51,9 @@ cc_test { "libnativewindow", ], static_libs: [ "VtsHalNeuralNetworksV1_0_utils", "VtsHalNeuralNetworksV1_2Callbacks", "VtsHalNeuralNetworksV1_3_utils", "android.hardware.neuralnetworks@1.0", "android.hardware.neuralnetworks@1.1", "android.hardware.neuralnetworks@1.2", Loading @@ -60,9 +64,6 @@ cc_test { "libhidlmemory", "libneuralnetworks_generated_test_harness", "libneuralnetworks_utils", "VtsHalNeuralNetworksV1_0_utils", "VtsHalNeuralNetworksV1_2Callbacks", "VtsHalNeuralNetworksV1_3Callbacks", ], whole_static_libs: [ "neuralnetworks_generated_V1_0_example", Loading
neuralnetworks/1.3/vts/functional/BasicTests.cpp +0 −1 Original line number Diff line number Diff line Loading @@ -21,7 +21,6 @@ namespace android::hardware::neuralnetworks::V1_3::vts::functional { using V1_0::DeviceStatus; using V1_0::ErrorStatus; using V1_0::PerformanceInfo; using V1_2::Constant; using V1_2::DeviceType; Loading