Home
last modified time | relevance | path

Searched refs:modelName (Results 1 – 23 of 23) sorted by relevance

/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/
H A Dnncompiler.cpp464 … std::string cacheInfo = cachePath + "/" + m_extensionConfig.modelName + "cache_info.nncache"; in OnlineBuild()
566 compiledCache.SetModelName(m_extensionConfig.modelName); in SaveToCacheFile()
604 compiledCache.SetModelName(m_extensionConfig.modelName); in RestoreFromCacheFile()
658 …std::string cacheInfoPath = m_cachePath + "/" + m_extensionConfig.modelName + "cache_info.nncache"; in RestoreFromCacheFile()
726 m_extensionConfig.modelName.assign(value.data(), value.data() + value.size()); in SetExtensionConfig()
727 …"[NNCompiler] SetExtensionConfig get model name:%{public}s.", m_extensionConfig.modelName.c_str()); in SetExtensionConfig()
742 OH_NN_ReturnCode NNCompiler::GetModelName(std::string& modelName) in GetModelName() argument
744 modelName = m_extensionConfig.modelName; in GetModelName()
H A Dneural_network_runtime.cpp442 extensionConfig.modelName.assign(value[0].first, value[0].first + value[0].second); in CheckExtensionConfigs()
570 NNRT_API bool OH_NNModel_HasCache(const char *cacheDir, const char *modelName, uint32_t version) in OH_NNModel_HasCache() argument
577 if (modelName == nullptr) { in OH_NNModel_HasCache()
582 …std::string cacheInfoPath = std::string(cacheDir) + "/" + std::string(modelName) + "cache_info.nnc… in OH_NNModel_HasCache()
608 std::string(cacheDir) + "/" + std::string(modelName) + std::to_string(i) + ".nncache"; in OH_NNModel_HasCache()
646 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz… in OH_NNModel_BuildFromMetaGraph()
H A Dnncompiled_cache.h53 void SetModelName(const std::string& modelName);
H A Dnncompiler.h54 OH_NN_ReturnCode GetModelName(std::string& modelName) override;
H A Dnncompiled_cache.cpp158 void NNCompiledCache::SetModelName(const std::string& modelName) in SetModelName() argument
160 m_modelName = modelName; in SetModelName()
/ohos5.0/foundation/ai/neural_network_runtime/example/deep_learning_framework/tflite/label_classify/
H A Dlabel_classify.cpp197 if (!settings.modelName.c_str()) { in InferenceModel()
203 model = tflite::FlatBufferModel::BuildFromFile(settings.modelName.c_str()); in InferenceModel()
205 LOG(ERROR) << "Failed to mmap model " << settings.modelName; in InferenceModel()
291 settings.modelName = optarg; in InitSettings()
H A Dlabel_classify.h36 string modelName = "./mbv2.tflite"; member
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/innerkits/c/
H A Dneural_network_runtime_inner.h142 bool OH_NNModel_HasCache(const char *cacheDir, const char *modelName, uint32_t version);
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/
H A Dcompiler.h49 virtual OH_NN_ReturnCode GetModelName(std::string& modelName) = 0;
H A Dnnrt_client.h33 size_t (*GetNNRtModelIDFromCache)(const char* path, const char* modelName) = nullptr;
H A Dcpp_type.h51 std::string modelName; member
H A Dneural_network_core.cpp624 std::string modelName; in GetNnrtModelId() local
625 OH_NN_ReturnCode retCode = compilationImpl->compiler->GetModelName(modelName); in GetNnrtModelId()
632 modelName.c_str()); in GetNnrtModelId()
640 nnrtService.GetNNRtModelIDFromCache(compilationImpl->cachePath, modelName.c_str()); in GetNnrtModelId()
/ohos5.0/docs/zh-cn/application-dev/ai/mindspore/
H A Dmindspore-guidelines-based-native.md159 …void *ReadModelFile(NativeResourceManager *nativeResourceManager, const std::string &modelName, si…
160 auto rawFile = OH_ResourceManager_OpenRawFile(nativeResourceManager, modelName.c_str());
320 const std::string modelName = "mobilenetv2.ms";
321 LOGI("MS_LITE_LOG: Run model: %{public}s", modelName.c_str());
324 auto modelBuffer = ReadModelFile(resourcesManager, modelName, &modelSize);
H A Dmindspore-guidelines-based-js.md216 let modelName: string = 'mobilenetv2.ms';
225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
/ohos5.0/docs/en/application-dev/ai/mindspore/
H A Dmindspore-guidelines-based-native.md159 …void *ReadModelFile(NativeResourceManager *nativeResourceManager, const std::string &modelName, si…
160 auto rawFile = OH_ResourceManager_OpenRawFile(nativeResourceManager, modelName.c_str());
320 const std::string modelName = "mobilenetv2.ms";
321 LOGI("MS_LITE_LOG: Run model: %{public}s", modelName.c_str());
324 auto modelBuffer = ReadModelFile(resourcesManager, modelName, &modelSize);
H A Dmindspore-guidelines-based-js.md216 let modelName: string = 'mobilenetv2.ms';
225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
/ohos5.0/base/update/updater/services/fs_manager/
H A Dpartitions.cpp199 static int InitGeneric(BlockDevice &dev, const std::string modelName) in InitGeneric() argument
220 dev.model = modelName; in InitGeneric()
/ohos5.0/base/security/security_guard/frameworks/js/napi/
H A Dsecurity_guard_napi.h100 std::string modelName; member
H A Dsecurity_guard_napi.cpp494 char modelName[MODEL_NAME_MAX_LEN] = {0}; in ParseModelRule() local
496 if (ParseString(env, napiValue, "modelName", modelName, len) == nullptr) { in ParseModelRule()
502 modelRule.modelName = std::string(modelName); in ParseModelRule()
521 if (ParseModelId(env, modelRule.modelName, modelId) == nullptr) { in NapiGetModelResult()
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/inner_model/
H A Dinner_model_test.cpp281 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz…
/ohos5.0/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/inner_model/
H A Dinner_model_test.cpp224 …extensionConfig.modelName.assign(extensions[i].value, extensions[i].value + extensions[i].valueSiz…
/ohos5.0/docs/zh-cn/application-dev/reference/apis-avsession-kit/
H A Djs-apis-avsession.md5270 | modelName<sup>13+</sup> | string | 否 | 播放设备型号名称。 <br> **系统能力:** SystemCapability.Multimedia.AVS…
/ohos5.0/docs/en/application-dev/reference/apis-avsession-kit/
H A Djs-apis-avsession.md5270 | modelName<sup>13+</sup> | string | No | Model name of the output device.<br> **System capability…