/ohos5.0/docs/zh-cn/device-dev/reference/hdi-apis/nnrt/ |
H A D | _model_config_v10.md | 20 | boolean [enableFloat16](#enablefloat16) | float32浮点模型是否以float16浮点运行 | 28 ### enableFloat16 subsection 31 boolean ModelConfig::enableFloat16
|
H A D | _model_config_v20.md | 20 | boolean [enableFloat16](#enablefloat16) | float32浮点模型是否以float16浮点运行 | 29 ### enableFloat16 subsection 32 boolean ModelConfig::enableFloat16
|
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/ |
H A D | hdi_device_v1_0.cpp | 267 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel() 319 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache()
|
H A D | hdi_device_v2_0.cpp | 310 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel() 361 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache() 579 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareOfflineModel()
|
H A D | hdi_device_v2_1.cpp | 305 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModel() 356 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareModelFromModelCache() 574 iModelConfig.enableFloat16 = config.enableFloat16; in PrepareOfflineModel()
|
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v2_0/cpp_target/nnrt/v2_0/ |
H A D | nnrt_types.cpp.txt | 159 if (!data.WriteBool(dataBlock.enableFloat16)) { 160 HDF_LOGE("%{public}s: write dataBlock.enableFloat16 failed!", __func__); 193 if (!data.ReadBool(dataBlock.enableFloat16)) { 194 HDF_LOGE("%{public}s: read dataBlock.enableFloat16 failed!", __func__);
|
H A D | nnrt_types.h.txt | 94 bool enableFloat16;
|
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/ |
H A D | cpp_type.h | 62 bool enableFloat16; member
|
H A D | neural_network_core.cpp | 384 … OH_NN_ReturnCode OH_NNCompilation_EnableFloat16(OH_NNCompilation *compilation, bool enableFloat16) in OH_NNCompilation_EnableFloat16() argument 392 compilationImpl->enableFp16 = enableFloat16; in OH_NNCompilation_EnableFloat16()
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/common/v1_0/ |
H A D | compilation_mock_idevice.cpp | 149 if (config.enableFloat16 == false) { in PrepareModel() 233 if (modelCache.size() == 0 || config.enableFloat16 == false) { in PrepareModelFromModelCache()
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/common/v2_0/ |
H A D | compilation_mock_idevice.cpp | 149 if (config.enableFloat16 == false) { in PrepareModel() 233 if (modelCache.size() == 0 || config.enableFloat16 == false) { in PrepareModelFromModelCache()
|
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/kits/c/neural_network_runtime/ |
H A D | neural_network_core.h | 323 OH_NN_ReturnCode OH_NNCompilation_EnableFloat16(OH_NNCompilation *compilation, bool enableFloat16);
|
/ohos5.0/drivers/interface/nnrt/v1_0/ |
H A D | NnrtTypes.idl | 55 boolean enableFloat16;
|
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/foo/nnrt/v1_0/ |
H A D | NnrtTypes.idl | 55 boolean enableFloat16;
|
/ohos5.0/drivers/interface/nnrt/v2_0/ |
H A D | NnrtTypes.idl | 55 boolean enableFloat16;
|
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v2_0/foo/nnrt/v2_0/ |
H A D | NnrtTypes.idl | 55 boolean enableFloat16;
|
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/cpp_target/nnrt/v1_0/ |
H A D | nnrt_types.h.txt | 93 bool enableFloat16;
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/neural_network_runtime_test/ |
H A D | neural_network_runtime_test.cpp | 34 if (config.enableFloat16 == false) { in PrepareModel() 976 bool enableFloat16 = true; variable 978 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16); 994 bool enableFloat16 = true; variable 996 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/ |
H A D | nnrt_device_service.cpp | 430 cpuInfo->SetEnableFP16(config.enableFloat16); in TransModelConfig()
|
/ohos5.0/docs/zh-cn/application-dev/reference/apis-neural-network-runtime-kit/ |
H A D | neural__network__core_8h.md | 39 …ion](_neural_network_runtime.md#oh_nncompilation) \*compilation, bool enableFloat16) | 是否以float16的…
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/neural_network_core_test/ |
H A D | neural_network_core_test.cpp | 850 bool enableFloat16 = true; variable 851 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16); 864 bool enableFloat16 = true; variable 865 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/src/ |
H A D | nnrt_device_service.cpp | 478 cpuInfo->SetEnableFP16(config.enableFloat16); in TransModelConfig()
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/neural_network_runtime_test/ |
H A D | neural_network_runtime_test.cpp | 35 if (config.enableFloat16 == false) { in PrepareModel() 1029 bool enableFloat16 = true; variable 1031 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16); 1047 bool enableFloat16 = true; variable 1049 OH_NN_ReturnCode ret = OH_NNCompilation_EnableFloat16(nnCompilation, enableFloat16);
|
/ohos5.0/docs/en/application-dev/reference/apis-neural-network-runtime-kit/ |
H A D | neural__network__core_8h.md | 39 …ion](_neural_network_runtime.md#oh_nncompilation) \*compilation, bool enableFloat16) | Enables flo…
|
/ohos5.0/foundation/ability/idl_tool/test/hdi_unittest/nnrt_v1_0/dump_ast_target/ |
H A D | dump.txt | 411 boolean enableFloat16;
|