/ohos5.0/docs/zh-cn/application-dev/ai/mindspore/ |
H A D | mindspore-guidelines-based-native.md | 166 void *modelBuffer = malloc(fileSize); 167 if (modelBuffer == nullptr) { 170 int ret = OH_ResourceManager_ReadRawFile(rawFile, modelBuffer, fileSize); 178 return modelBuffer; 193 OH_AI_ContextHandle CreateMSLiteContext(void *modelBuffer) { 197 DestroyModelBuffer(&modelBuffer); 214 DestroyModelBuffer(&modelBuffer); 221 DestroyModelBuffer(&modelBuffer); 325 if (modelBuffer == nullptr) { 331 auto context = CreateMSLiteContext(modelBuffer); [all …]
|
H A D | mindspore-guidelines-based-js.md | 179 modelBuffer: ArrayBuffer, inputsBuffer: ArrayBuffer[]): Promise<mindSporeLite.MSTensor[]> { 190 …let msLiteModel: mindSporeLite.Model = await mindSporeLite.loadModelFromBuffer(modelBuffer, contex… 225 resMgr.getRawFileContent(modelName).then(modelBuffer => { 227 modelPredict(modelBuffer.buffer.slice(0), inputs).then(outputs => {
|
/ohos5.0/docs/zh-cn/application-dev/reference/apis-mindspore-lite-kit/ |
H A D | js-apis-mindSporeLite.md | 123 let modelBuffer = buffer.buffer; 163 let modelBuffer = buffer.buffer; 210 let modelBuffer = buffer.buffer; 365 resMgr.getRawFileContent(modelFile).then(modelBuffer => { 769 let modelBuffer = buffer.buffer; 774 modelInputs[0].setData(modelBuffer); 819 let modelBuffer = buffer.buffer; 823 modelInputs[0].setData(modelBuffer); 920 resMgr.getRawFileContent(modelFile).then((modelBuffer) => { 963 resMgr.getRawFileContent(modelFile).then((modelBuffer) => { [all …]
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/ |
H A D | prepared_model_service.cpp | 203 uint8_t* modelBuffer = m_builder.GetBufferPointer(); in Compile() local 204 if (modelBuffer == nullptr) { in Compile() 210 mindspore::Status msRet = m_model->Build(modelBuffer, modelSize, mindspore::kMindIR, m_context); in Compile() 224 int32_t PreparedModelService::Compile(const void* modelBuffer, size_t length) in Compile() argument 226 if (modelBuffer == nullptr || length == 0) { in Compile() 232 mindspore::Status msRet = m_model->Build(modelBuffer, length, mindspore::kMindIR, m_context); in Compile()
|
H A D | nnrt_device_service.cpp | 167 void* modelBuffer = parser.GetBufferPtr(); in PrepareModelFromModelCache() local 175 ret = service->Compile(modelBuffer, modelCache[0].dataSize); in PrepareModelFromModelCache()
|
/ohos5.0/docs/en/application-dev/ai/mindspore/ |
H A D | mindspore-guidelines-based-native.md | 166 void *modelBuffer = malloc(fileSize); 167 if (modelBuffer == nullptr) { 170 int ret = OH_ResourceManager_ReadRawFile(rawFile, modelBuffer, fileSize); 178 return modelBuffer; 193 OH_AI_ContextHandle CreateMSLiteContext(void *modelBuffer) { 197 DestroyModelBuffer(&modelBuffer); 214 DestroyModelBuffer(&modelBuffer); 221 DestroyModelBuffer(&modelBuffer); 325 if (modelBuffer == nullptr) { 331 auto context = CreateMSLiteContext(modelBuffer); [all …]
|
H A D | mindspore-guidelines-based-js.md | 179 modelBuffer: ArrayBuffer, inputsBuffer: ArrayBuffer[]): Promise<mindSporeLite.MSTensor[]> { 190 …let msLiteModel: mindSporeLite.Model = await mindSporeLite.loadModelFromBuffer(modelBuffer, contex… 225 resMgr.getRawFileContent(modelName).then(modelBuffer => { 227 modelPredict(modelBuffer.buffer.slice(0), inputs).then(outputs => {
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/src/ |
H A D | prepared_model_service.cpp | 243 uint8_t* modelBuffer = m_builder.GetBufferPointer(); in Compile() local 244 if (modelBuffer == nullptr) { in Compile() 250 mindspore::Status msRet = m_model->Build(modelBuffer, modelSize, mindspore::kMindIR, m_context); in Compile() 269 NNRT_ReturnCode PreparedModelService::Compile(const void* modelBuffer, size_t length) in Compile() argument 271 if (modelBuffer == nullptr || length == 0) { in Compile() 277 mindspore::Status msRet = m_model->Build(modelBuffer, length, mindspore::kMindIR, m_context); in Compile()
|
H A D | nnrt_device_service.cpp | 208 void* modelBuffer = parser.GetBufferPtr(); in PrepareModelFromModelCache() local 209 ret = service->Compile(modelBuffer, modelCache[0].dataSize); in PrepareModelFromModelCache()
|
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/ |
H A D | nncompiled_cache.cpp | 118 OHOS::NeuralNetworkRuntime::Buffer modelBuffer; in Restore() local 119 ret = ReadCacheModelFile(cacheModelPath, modelBuffer); in Restore() 125 if (GetCrc16(static_cast<char*>(modelBuffer.data), modelBuffer.length) != in Restore() 132 caches.emplace_back(std::move(modelBuffer)); in Restore()
|
H A D | hdi_device_v2_0.cpp | 533 Buffer modelBuffer {nullptr, 0}; in AllocateDeviceBufferForOfflineModel() local 534 modelBuffer.data = newModelBuffer; in AllocateDeviceBufferForOfflineModel() 535 modelBuffer.length = offlineModelSize; in AllocateDeviceBufferForOfflineModel() 536 deviceBuffers.emplace_back(modelBuffer); in AllocateDeviceBufferForOfflineModel()
|
H A D | hdi_device_v2_1.cpp | 528 Buffer modelBuffer {nullptr, 0}; in AllocateDeviceBufferForOfflineModel() local 529 modelBuffer.data = newModelBuffer; in AllocateDeviceBufferForOfflineModel() 530 modelBuffer.length = offlineModelSize; in AllocateDeviceBufferForOfflineModel() 531 deviceBuffers.emplace_back(modelBuffer); in AllocateDeviceBufferForOfflineModel()
|
/ohos5.0/docs/en/application-dev/reference/apis-mindspore-lite-kit/ |
H A D | js-apis-mindSporeLite.md | 123 let modelBuffer = buffer.buffer; 163 let modelBuffer = buffer.buffer; 210 let modelBuffer = buffer.buffer; 365 resMgr.getRawFileContent(modelFile).then(modelBuffer => { 769 let modelBuffer = buffer.buffer; 774 modelInputs[0].setData(modelBuffer); 819 let modelBuffer = buffer.buffer; 823 modelInputs[0].setData(modelBuffer); 920 resMgr.getRawFileContent(modelFile).then((modelBuffer) => { 963 resMgr.getRawFileContent(modelFile).then((modelBuffer) => { [all …]
|
/ohos5.0/foundation/ai/neural_network_runtime/test/fuzztest/ |
H A D | nnrt_device_impl.h | 90 …int32_t PrepareOfflineModel(const std::vector<SharedBuffer>& modelBuffer, const ModelConfig& confi… in PrepareOfflineModel() argument
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/include/ |
H A D | prepared_model_service.h | 42 int32_t Compile(const void* modelBuffer, size_t length);
|
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/include/ |
H A D | prepared_model_service.h | 42 NNRT_ReturnCode Compile(const void* modelBuffer, size_t length);
|
H A D | nnrt_device_service.h | 56 …int32_t PrepareOfflineModel(const std::vector<SharedBuffer>& modelBuffer, const ModelConfig& confi…
|
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/kits/c/neural_network_runtime/ |
H A D | neural_network_core.h | 111 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_t …
|
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/ |
H A D | neural_network_core.cpp | 153 NNRT_API OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer… in OH_NNCompilation_ConstructWithOfflineModelBuffer() argument 155 if (modelBuffer == nullptr) { in OH_NNCompilation_ConstructWithOfflineModelBuffer() 171 compilation->offlineModelBuffer.first = const_cast<void*>(modelBuffer); in OH_NNCompilation_ConstructWithOfflineModelBuffer()
|
/ohos5.0/docs/zh-cn/application-dev/reference/apis-neural-network-runtime-kit/ |
H A D | neural__network__core_8h.md | 30 …ime.md#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
|
H A D | _neural_network_runtime.md | 81 …ffer](#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi… 1060 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer (const void *modelBuffer, size_t… 1069 …tion](#oh_nncompilation)实例只将**modelBuffer**指针保存在里面,而不是复制其数据。 在销毁[OH_NNCompilation](#oh_nncompilati… 1077 | modelBuffer | 离线模型文件内存。 |
|
H A D | _neural_nework_runtime.md | 81 …ffer](#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi… 1060 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer (const void *modelBuffer, size_t… 1069 …tion](#oh_nncompilation)实例只将**modelBuffer**指针保存在里面,而不是复制其数据。 在销毁[OH_NNCompilation](#oh_nncompilati… 1077 | modelBuffer | 离线模型文件内存。 |
|
/ohos5.0/docs/en/application-dev/reference/apis-neural-network-runtime-kit/ |
H A D | neural__network__core_8h.md | 30 …ime.md#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
|
/ohos5.0/foundation/ai/neural_network_runtime/ |
H A D | neural-network-runtime-guidelines.md | 79 | OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_…
|
/ohos5.0/docs/zh-cn/application-dev/ai/nnrt/ |
H A D | neural-network-runtime-guidelines.md | 79 | OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_…
|