Loading neuralnetworks/1.0/utils/include/nnapi/hal/1.0/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -50,8 +50,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading neuralnetworks/1.0/utils/src/PreparedModel.cpp +4 −5 Original line number Diff line number Diff line Loading @@ -63,12 +63,11 @@ nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> Prepare // Ensure that request is ready for IPC. std::optional<nn::Request> maybeRequestInShared; hal::utils::RequestRelocation relocation; const nn::Request& requestInShared = NN_TRY(hal::utils::makeExecutionFailure(hal::utils::convertRequestFromPointerToShared( const nn::Request& requestInShared = NN_TRY(hal::utils::convertRequestFromPointerToShared( &request, nn::kDefaultRequestMemoryAlignment, nn::kMinMemoryPadding, &maybeRequestInShared, &relocation))); &maybeRequestInShared, &relocation)); const auto hidlRequest = NN_TRY(hal::utils::makeExecutionFailure(convert(requestInShared))); const auto hidlRequest = NN_TRY(convert(requestInShared)); return executeInternal(hidlRequest, relocation); } Loading neuralnetworks/1.1/utils/include/nnapi/hal/1.1/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -52,8 +52,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading neuralnetworks/1.2/utils/include/nnapi/hal/1.2/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -61,8 +61,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading neuralnetworks/1.2/utils/src/Callbacks.cpp +1 −2 Original line number Diff line number Diff line Loading @@ -75,8 +75,7 @@ nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> executi << "execution failed with " << toString(status); } HANDLE_HAL_STATUS(status) << "execution failed with " << toString(status); return hal::utils::makeExecutionFailure( convertExecutionGeneralResultsHelper(outputShapes, timing)); return convertExecutionGeneralResultsHelper(outputShapes, timing); } Return<void> PreparedModelCallback::notify(V1_0::ErrorStatus status, Loading Loading
neuralnetworks/1.0/utils/include/nnapi/hal/1.0/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -50,8 +50,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading
neuralnetworks/1.0/utils/src/PreparedModel.cpp +4 −5 Original line number Diff line number Diff line Loading @@ -63,12 +63,11 @@ nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> Prepare // Ensure that request is ready for IPC. std::optional<nn::Request> maybeRequestInShared; hal::utils::RequestRelocation relocation; const nn::Request& requestInShared = NN_TRY(hal::utils::makeExecutionFailure(hal::utils::convertRequestFromPointerToShared( const nn::Request& requestInShared = NN_TRY(hal::utils::convertRequestFromPointerToShared( &request, nn::kDefaultRequestMemoryAlignment, nn::kMinMemoryPadding, &maybeRequestInShared, &relocation))); &maybeRequestInShared, &relocation)); const auto hidlRequest = NN_TRY(hal::utils::makeExecutionFailure(convert(requestInShared))); const auto hidlRequest = NN_TRY(convert(requestInShared)); return executeInternal(hidlRequest, relocation); } Loading
neuralnetworks/1.1/utils/include/nnapi/hal/1.1/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -52,8 +52,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading
neuralnetworks/1.2/utils/include/nnapi/hal/1.2/Utils.h +2 −2 Original line number Diff line number Diff line Loading @@ -61,8 +61,8 @@ bool valid(const Type& halObject) { } template <typename Type> nn::GeneralResult<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(hal::utils::makeGeneralFailure(nn::validate(canonical))); nn::Result<void> compliantVersion(const Type& canonical) { const auto version = NN_TRY(nn::validate(canonical)); if (version > kVersion) { return NN_ERROR() << "Insufficient version: " << version << " vs required " << kVersion; } Loading
neuralnetworks/1.2/utils/src/Callbacks.cpp +1 −2 Original line number Diff line number Diff line Loading @@ -75,8 +75,7 @@ nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> executi << "execution failed with " << toString(status); } HANDLE_HAL_STATUS(status) << "execution failed with " << toString(status); return hal::utils::makeExecutionFailure( convertExecutionGeneralResultsHelper(outputShapes, timing)); return convertExecutionGeneralResultsHelper(outputShapes, timing); } Return<void> PreparedModelCallback::notify(V1_0::ErrorStatus status, Loading