Home
last modified time | relevance | path

Searched refs:enableFloat16 (Results 1 – 25 of 34) sorted by relevance

12

/ohos5.0/docs/zh-cn/device-dev/reference/hdi-apis/nnrt/
H A D_model_config_v10.md20 | boolean [enableFloat16](#enablefloat16) | float32浮点模型是否以float16浮点运行 |
28 ### enableFloat16 subsection
31 boolean ModelConfig::enableFloat16
H A D_model_config_v20.md20 | boolean [enableFloat16](#enablefloat16) | float32浮点模型是否以float16浮点运行 |
29 ### enableFloat16 subsection
32 boolean ModelConfig::enableFloat16
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/
H A Dhdi_device_v1_0.cpp267 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel()
319 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache()
H A Dhdi_device_v2_0.cpp310 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel()
361 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache()
579 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareOfflineModel()
H A Dhdi_device_v2_1.cpp305 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel()
356 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache()
574 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareOfflineModel()
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v2_0/cpp_target/nnrt/v2_0/
H A Dnnrt_types.cpp.txt159 if (!data.WriteBool(dataBlock.enableFloat16)) {
160 HDF_LOGE("%{public}s: write dataBlock.enableFloat16 failed!", __func__);
193 if (!data.ReadBool(dataBlock.enableFloat16)) {
194 HDF_LOGE("%{public}s: read dataBlock.enableFloat16 failed!", __func__);
H A Dnnrt_types.h.txt94 bool enableFloat16;
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/
H A Dcpp_type.h62 bool enableFloat16; member
H A Dneural_network_core.cpp384 … OH_NN_ReturnCode OH_NNCompilation_EnableFloat16(OH_NNCompilation *compilation, bool enableFloat16) in OH_NNCompilation_EnableFloat16() argument
392 compilationImpl->enableFp16 = enableFloat16; in OH_NNCompilation_EnableFloat16()
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/common/v1_0/
H A Dcompilation_mock_idevice.cpp149 if (config.enableFloat16 == false) { in PrepareModel()
233 if (modelCache.size() == 0 || config.enableFloat16 == false) { in PrepareModelFromModelCache()
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/common/v2_0/
H A Dcompilation_mock_idevice.cpp149 if (config.enableFloat16 == false) { in PrepareModel()
233 if (modelCache.size() == 0 || config.enableFloat16 == false) { in PrepareModelFromModelCache()
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/kits/c/neural_network_runtime/
H A Dneural_network_core.h323 OH_NN_ReturnCode OH_NNCompilation_EnableFloat16(OH_NNCompilation *compilation, bool enableFloat16);
/ohos5.0/drivers/interface/nnrt/v1_0/
H A DNnrtTypes.idl55 boolean enableFloat16;
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/foo/nnrt/v1_0/
H A DNnrtTypes.idl55 boolean enableFloat16;
/ohos5.0/drivers/interface/nnrt/v2_0/
H A DNnrtTypes.idl55 boolean enableFloat16;
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v2_0/foo/nnrt/v2_0/
H A DNnrtTypes.idl55 boolean enableFloat16;
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/cpp_target/nnrt/v1_0/
H A Dnnrt_types.h.txt93 bool enableFloat16;
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/neural_network_runtime_test/
H A Dneural_network_runtime_test.cpp34 if (config.enableFloat16 == false) { in PrepareModel()
976 bool enableFloat16 = true; variable
978 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
994 bool enableFloat16 = true; variable
996 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/
H A Dnnrt_device_service.cpp430 cpuInfo->SetEnableFP16(config.enableFloat16); in TransModelConfig()
/ohos5.0/docs/zh-cn/application-dev/reference/apis-neural-network-runtime-kit/
H A Dneural__network__core_8h.md39 …ion](_neural_network_runtime.md#oh_nncompilation) \*compilation, bool enableFloat16) | 是否以float16的…
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/neural_network_core_test/
H A Dneural_network_core_test.cpp850 bool enableFloat16 = true; variable
851 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
864 bool enableFloat16 = true; variable
865 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/src/
H A Dnnrt_device_service.cpp478 cpuInfo->SetEnableFP16(config.enableFloat16); in TransModelConfig()
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/neural_network_runtime_test/
H A Dneural_network_runtime_test.cpp35 if (config.enableFloat16 == false) { in PrepareModel()
1029 bool enableFloat16 = true; variable
1031 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
1047 bool enableFloat16 = true; variable
1049 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
/ohos5.0/docs/en/application-dev/reference/apis-neural-network-runtime-kit/
H A Dneural__network__core_8h.md39 …ion](_neural_network_runtime.md#oh_nncompilation) \*compilation, bool enableFloat16) | Enables flo…
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/dump_ast_target/
H A Ddump.txt411 boolean enableFloat16;

12