/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/ |
H A D | nncompiler.cpp | 464 … std::string cacheInfo = cachePath + "/" + m_extensionConfig.modelName + "cache_info.nncache"; in OnlineBuild() 566 compiledCache.SetModelName(m_extensionConfig.modelName); in SaveToCacheFile() 604 compiledCache.SetModelName(m_extensionConfig.modelName); in RestoreFromCacheFile() 658 …std::string cacheInfoPath = m_cachePath + "/" + m_extensionConfig.modelName + "cache_info.nncache"; in RestoreFromCacheFile() 726 m_extensionConfig.modelName.assign(value.data(), value.data() + value.size()); in SetExtensionConfig() 727 …"[NNCompiler] SetExtensionConfig get model name:%{public}s.", m_extensionConfig.modelName.c_str()); in SetExtensionConfig() 742 OH_NN_ReturnCode NNCompiler::GetModelName(std::string& modelName) in GetModelName() argument 744 modelName = m_extensionConfig.modelName; in GetModelName()
|
H A D | neural_network_runtime.cpp | 442 extensionConfig.modelName.assign(value[0].first, value[0].first + value[0].second); in CheckExtensionConfigs() 570 NNRT_API bool OH_NNModel_HasCache(const char *cacheDir, const char *modelName, uint32_t version) in OH_NNModel_HasCache() argument 577 if (modelName == nullptr) { in OH_NNModel_HasCache() 582 …std::string cacheInfoPath = std::string(cacheDir) + "/" + std::string(modelName) + "cache_info.nnc… in OH_NNModel_HasCache() 608 std::string(cacheDir) + "/" + std::string(modelName) + std::to_string(i) + ".nncache"; in OH_NNModel_HasCache() 646 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz… in OH_NNModel_BuildFromMetaGraph()
|
H A D | nncompiled_cache.h | 53 void SetModelName(const std::string& modelName);
|
H A D | nncompiler.h | 54 OH_NN_ReturnCode GetModelName(std::string& modelName) override;
|
H A D | nncompiled_cache.cpp | 158 void NNCompiledCache::SetModelName(const std::string& modelName) in SetModelName() argument 160 m_modelName = modelName; in SetModelName()
|
/ohos5.0/foundation/ai/neural_network_runtime/example/deep_learning_framework/tflite/label_classify/ |
H A D | label_classify.cpp | 197 if (!settings.modelName.c_str()) { in InferenceModel() 203 model = tflite::FlatBufferModel::BuildFromFile(settings.modelName.c_str()); in InferenceModel() 205 LOG(ERROR) << "Failed to mmap model " << settings.modelName; in InferenceModel() 291 settings.modelName = optarg; in InitSettings()
|
H A D | label_classify.h | 36 string modelName = "./mbv2.tflite"; member
|
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/innerkits/c/ |
H A D | neural_network_runtime_inner.h | 142 bool OH_NNModel_HasCache(const char *cacheDir, const char *modelName, uint32_t version);
|
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/ |
H A D | compiler.h | 49 virtual OH_NN_ReturnCode GetModelName(std::string& modelName) = 0;
|
H A D | nnrt_client.h | 33 size_t (*GetNNRtModelIDFromCache)(const char* path, const char* modelName) = nullptr;
|
H A D | cpp_type.h | 51 std::string modelName; member
|
H A D | neural_network_core.cpp | 624 std::string modelName; in GetNnrtModelId() local 625 OH_NN_ReturnCode retCode = compilationImpl->compiler->GetModelName(modelName); in GetNnrtModelId() 632 modelName.c_str()); in GetNnrtModelId() 640 nnrtService.GetNNRtModelIDFromCache(compilationImpl->cachePath, modelName.c_str()); in GetNnrtModelId()
|
/ohos5.0/docs/zh-cn/application-dev/ai/mindspore/ |
H A D | mindspore-guidelines-based-native.md | 159 …void *ReadModelFile(NativeResourceManager *nativeResourceManager, const std::string &modelName, si… 160 auto rawFile = OH_ResourceManager_OpenRawFile(nativeResourceManager, modelName.c_str()); 320 const std::string modelName = "mobilenetv2.ms"; 321 LOGI("MS_LITE_LOG: Run model: %{public}s", modelName.c_str()); 324 auto modelBuffer = ReadModelFile(resourcesManager, modelName, &modelSize);
|
H A D | mindspore-guidelines-based-js.md | 216 let modelName: string = 'mobilenetv2.ms'; 225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
|
/ohos5.0/docs/en/application-dev/ai/mindspore/ |
H A D | mindspore-guidelines-based-native.md | 159 …void *ReadModelFile(NativeResourceManager *nativeResourceManager, const std::string &modelName, si… 160 auto rawFile = OH_ResourceManager_OpenRawFile(nativeResourceManager, modelName.c_str()); 320 const std::string modelName = "mobilenetv2.ms"; 321 LOGI("MS_LITE_LOG: Run model: %{public}s", modelName.c_str()); 324 auto modelBuffer = ReadModelFile(resourcesManager, modelName, &modelSize);
|
H A D | mindspore-guidelines-based-js.md | 216 let modelName: string = 'mobilenetv2.ms'; 225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
|
/ohos5.0/base/update/updater/services/fs_manager/ |
H A D | partitions.cpp | 199 static int InitGeneric(BlockDevice &dev, const std::string modelName) in InitGeneric() argument 220 dev.model = modelName; in InitGeneric()
|
/ohos5.0/base/security/security_guard/frameworks/js/napi/ |
H A D | security_guard_napi.h | 100 std::string modelName; member
|
H A D | security_guard_napi.cpp | 494 char modelName[MODEL_NAME_MAX_LEN] = {0}; in ParseModelRule() local 496 if (ParseString(env, napiValue, "modelName", modelName, len) == nullptr) { in ParseModelRule() 502 modelRule.modelName = std::string(modelName); in ParseModelRule() 521 if (ParseModelId(env, modelRule.modelName, modelId) == nullptr) { in NapiGetModelResult()
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/inner_model/ |
H A D | inner_model_test.cpp | 281 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz…
|
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/inner_model/ |
H A D | inner_model_test.cpp | 224 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz…
|
/ohos5.0/docs/zh-cn/application-dev/reference/apis-avsession-kit/ |
H A D | js-apis-avsession.md | 5270 | modelName<sup>13+</sup> | string | 否 | 播放设备型号名称。 <br> **系统能力:** SystemCapability.Multimedia.AVS…
|
/ohos5.0/docs/en/application-dev/reference/apis-avsession-kit/ |
H A D | js-apis-avsession.md | 5270 | modelName<sup>13+</sup> | string | No | Model name of the output device.<br> **System capability…
|