Loading neuralnetworks/1.3/utils/src/Conversions.cpp +2 −2 Original line number Diff line number Diff line Loading @@ -244,7 +244,7 @@ GeneralResult<BufferRole> unvalidatedConvert(const hal::V1_3::BufferRole& buffer return BufferRole{ .modelIndex = bufferRole.modelIndex, .ioIndex = bufferRole.ioIndex, .frequency = bufferRole.frequency, .probability = bufferRole.frequency, }; } Loading Loading @@ -577,7 +577,7 @@ nn::GeneralResult<BufferRole> unvalidatedConvert(const nn::BufferRole& bufferRol return BufferRole{ .modelIndex = bufferRole.modelIndex, .ioIndex = bufferRole.ioIndex, .frequency = bufferRole.frequency, .frequency = bufferRole.probability, }; } Loading neuralnetworks/aidl/aidl_api/android.hardware.neuralnetworks/current/android/hardware/neuralnetworks/BufferRole.aidl +1 −1 Original line number Diff line number Diff line Loading @@ -36,5 +36,5 @@ package android.hardware.neuralnetworks; parcelable BufferRole { int modelIndex; int ioIndex; float frequency; float probability; } neuralnetworks/aidl/android/hardware/neuralnetworks/BufferRole.aidl +1 −1 Original line number Diff line number Diff line Loading @@ -35,5 +35,5 @@ parcelable BufferRole { * used in the specified role. This is provided as a hint to optimize the case when multiple * roles prefer different buffer locations or data layouts. */ float frequency; float probability; } neuralnetworks/aidl/utils/src/Conversions.cpp +2 −2 Original line number Diff line number Diff line Loading @@ -472,7 +472,7 @@ GeneralResult<BufferRole> unvalidatedConvert(const aidl_hal::BufferRole& bufferR return BufferRole{ .modelIndex = static_cast<uint32_t>(bufferRole.modelIndex), .ioIndex = static_cast<uint32_t>(bufferRole.ioIndex), .frequency = bufferRole.frequency, .probability = bufferRole.probability, }; } Loading Loading @@ -718,7 +718,7 @@ nn::GeneralResult<BufferRole> unvalidatedConvert(const nn::BufferRole& bufferRol return BufferRole{ .modelIndex = static_cast<int32_t>(bufferRole.modelIndex), .ioIndex = static_cast<int32_t>(bufferRole.ioIndex), .frequency = bufferRole.frequency, .probability = bufferRole.probability, }; } Loading neuralnetworks/aidl/vts/functional/GeneratedTestHarness.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -102,7 +102,7 @@ class DeviceMemoryAllocator { ASSERT_NE(result, nullptr); // Prepare arguments. BufferRole role = {.modelIndex = 0, .ioIndex = index, .frequency = 1.0f}; BufferRole role = {.modelIndex = 0, .ioIndex = index, .probability = 1.0f}; std::vector<BufferRole> inputRoles, outputRoles; if constexpr (ioType == IOType::INPUT) { inputRoles = {role}; Loading Loading
neuralnetworks/1.3/utils/src/Conversions.cpp +2 −2 Original line number Diff line number Diff line Loading @@ -244,7 +244,7 @@ GeneralResult<BufferRole> unvalidatedConvert(const hal::V1_3::BufferRole& buffer return BufferRole{ .modelIndex = bufferRole.modelIndex, .ioIndex = bufferRole.ioIndex, .frequency = bufferRole.frequency, .probability = bufferRole.frequency, }; } Loading Loading @@ -577,7 +577,7 @@ nn::GeneralResult<BufferRole> unvalidatedConvert(const nn::BufferRole& bufferRol return BufferRole{ .modelIndex = bufferRole.modelIndex, .ioIndex = bufferRole.ioIndex, .frequency = bufferRole.frequency, .frequency = bufferRole.probability, }; } Loading
neuralnetworks/aidl/aidl_api/android.hardware.neuralnetworks/current/android/hardware/neuralnetworks/BufferRole.aidl +1 −1 Original line number Diff line number Diff line Loading @@ -36,5 +36,5 @@ package android.hardware.neuralnetworks; parcelable BufferRole { int modelIndex; int ioIndex; float frequency; float probability; }
neuralnetworks/aidl/android/hardware/neuralnetworks/BufferRole.aidl +1 −1 Original line number Diff line number Diff line Loading @@ -35,5 +35,5 @@ parcelable BufferRole { * used in the specified role. This is provided as a hint to optimize the case when multiple * roles prefer different buffer locations or data layouts. */ float frequency; float probability; }
neuralnetworks/aidl/utils/src/Conversions.cpp +2 −2 Original line number Diff line number Diff line Loading @@ -472,7 +472,7 @@ GeneralResult<BufferRole> unvalidatedConvert(const aidl_hal::BufferRole& bufferR return BufferRole{ .modelIndex = static_cast<uint32_t>(bufferRole.modelIndex), .ioIndex = static_cast<uint32_t>(bufferRole.ioIndex), .frequency = bufferRole.frequency, .probability = bufferRole.probability, }; } Loading Loading @@ -718,7 +718,7 @@ nn::GeneralResult<BufferRole> unvalidatedConvert(const nn::BufferRole& bufferRol return BufferRole{ .modelIndex = static_cast<int32_t>(bufferRole.modelIndex), .ioIndex = static_cast<int32_t>(bufferRole.ioIndex), .frequency = bufferRole.frequency, .probability = bufferRole.probability, }; } Loading
neuralnetworks/aidl/vts/functional/GeneratedTestHarness.cpp +1 −1 Original line number Diff line number Diff line Loading @@ -102,7 +102,7 @@ class DeviceMemoryAllocator { ASSERT_NE(result, nullptr); // Prepare arguments. BufferRole role = {.modelIndex = 0, .ioIndex = index, .frequency = 1.0f}; BufferRole role = {.modelIndex = 0, .ioIndex = index, .probability = 1.0f}; std::vector<BufferRole> inputRoles, outputRoles; if constexpr (ioType == IOType::INPUT) { inputRoles = {role}; Loading