Home
last modified time | relevance | path

Searched refs:modelBuffer (Results 1 – 25 of 27) sorted by relevance

12

/ohos5.0/docs/zh-cn/application-dev/ai/mindspore/
H A Dmindspore-guidelines-based-native.md166 void *modelBuffer = malloc(fileSize);
167 if (modelBuffer == nullptr) {
170 int ret = OH_ResourceManager_ReadRawFile(rawFile, modelBuffer, fileSize);
178 return modelBuffer;
193 OH_AI_ContextHandle CreateMSLiteContext(void *modelBuffer) {
197 DestroyModelBuffer(&modelBuffer);
214 DestroyModelBuffer(&modelBuffer);
221 DestroyModelBuffer(&modelBuffer);
325 if (modelBuffer == nullptr) {
331 auto context = CreateMSLiteContext(modelBuffer);
[all …]
H A Dmindspore-guidelines-based-js.md179 modelBuffer: ArrayBuffer, inputsBuffer: ArrayBuffer[]): Promise<mindSporeLite.MSTensor[]> {
190 …let msLiteModel: mindSporeLite.Model = await mindSporeLite.loadModelFromBuffer(modelBuffer, contex…
225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
227 modelPredict(modelBuffer.buffer.slice(0), inputs).then(outputs => {
/ohos5.0/docs/zh-cn/application-dev/reference/apis-mindspore-lite-kit/
H A Djs-apis-mindSporeLite.md123 let modelBuffer = buffer.buffer;
163 let modelBuffer = buffer.buffer;
210 let modelBuffer = buffer.buffer;
365 resMgr.getRawFileContent(modelFile).then(modelBuffer => {
769 let modelBuffer = buffer.buffer;
774 modelInputs[0].setData(modelBuffer);
819 let modelBuffer = buffer.buffer;
823 modelInputs[0].setData(modelBuffer);
920 resMgr.getRawFileContent(modelFile).then((modelBuffer) => {
963 resMgr.getRawFileContent(modelFile).then((modelBuffer) => {
[all …]
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/
H A Dprepared_model_service.cpp203 uint8_t* modelBuffer = m_builder.GetBufferPointer(); in Compile() local
204 if (modelBuffer == nullptr) { in Compile()
210 mindspore::Status msRet = m_model->Build(modelBuffer, modelSize, mindspore::kMindIR, m_context); in Compile()
224 int32_t PreparedModelService::Compile(const void* modelBuffer, size_t length) in Compile() argument
226 if (modelBuffer == nullptr || length == 0) { in Compile()
232 mindspore::Status msRet = m_model->Build(modelBuffer, length, mindspore::kMindIR, m_context); in Compile()
H A Dnnrt_device_service.cpp167 void* modelBuffer = parser.GetBufferPtr(); in PrepareModelFromModelCache() local
175 ret = service->Compile(modelBuffer, modelCache[0].dataSize); in PrepareModelFromModelCache()
/ohos5.0/docs/en/application-dev/ai/mindspore/
H A Dmindspore-guidelines-based-native.md166 void *modelBuffer = malloc(fileSize);
167 if (modelBuffer == nullptr) {
170 int ret = OH_ResourceManager_ReadRawFile(rawFile, modelBuffer, fileSize);
178 return modelBuffer;
193 OH_AI_ContextHandle CreateMSLiteContext(void *modelBuffer) {
197 DestroyModelBuffer(&modelBuffer);
214 DestroyModelBuffer(&modelBuffer);
221 DestroyModelBuffer(&modelBuffer);
325 if (modelBuffer == nullptr) {
331 auto context = CreateMSLiteContext(modelBuffer);
[all …]
H A Dmindspore-guidelines-based-js.md179 modelBuffer: ArrayBuffer, inputsBuffer: ArrayBuffer[]): Promise<mindSporeLite.MSTensor[]> {
190 …let msLiteModel: mindSporeLite.Model = await mindSporeLite.loadModelFromBuffer(modelBuffer, contex…
225 resMgr.getRawFileContent(modelName).then(modelBuffer => {
227 modelPredict(modelBuffer.buffer.slice(0), inputs).then(outputs => {
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/src/
H A Dprepared_model_service.cpp243 uint8_t* modelBuffer = m_builder.GetBufferPointer(); in Compile() local
244 if (modelBuffer == nullptr) { in Compile()
250 mindspore::Status msRet = m_model->Build(modelBuffer, modelSize, mindspore::kMindIR, m_context); in Compile()
269 NNRT_ReturnCode PreparedModelService::Compile(const void* modelBuffer, size_t length) in Compile() argument
271 if (modelBuffer == nullptr || length == 0) { in Compile()
277 mindspore::Status msRet = m_model->Build(modelBuffer, length, mindspore::kMindIR, m_context); in Compile()
H A Dnnrt_device_service.cpp208 void* modelBuffer = parser.GetBufferPtr(); in PrepareModelFromModelCache() local
209 ret = service->Compile(modelBuffer, modelCache[0].dataSize); in PrepareModelFromModelCache()
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/
H A Dnncompiled_cache.cpp118 OHOS::NeuralNetworkRuntime::Buffer modelBuffer; in Restore() local
119 ret = ReadCacheModelFile(cacheModelPath, modelBuffer); in Restore()
125 if (GetCrc16(static_cast<char*>(modelBuffer.data), modelBuffer.length) != in Restore()
132 caches.emplace_back(std::move(modelBuffer)); in Restore()
H A Dhdi_device_v2_0.cpp533 Buffer modelBuffer {nullptr, 0}; in AllocateDeviceBufferForOfflineModel() local
534 modelBuffer.data = newModelBuffer; in AllocateDeviceBufferForOfflineModel()
535 modelBuffer.length = offlineModelSize; in AllocateDeviceBufferForOfflineModel()
536 deviceBuffers.emplace_back(modelBuffer); in AllocateDeviceBufferForOfflineModel()
H A Dhdi_device_v2_1.cpp528 Buffer modelBuffer {nullptr, 0}; in AllocateDeviceBufferForOfflineModel() local
529 modelBuffer.data = newModelBuffer; in AllocateDeviceBufferForOfflineModel()
530 modelBuffer.length = offlineModelSize; in AllocateDeviceBufferForOfflineModel()
531 deviceBuffers.emplace_back(modelBuffer); in AllocateDeviceBufferForOfflineModel()
/ohos5.0/docs/en/application-dev/reference/apis-mindspore-lite-kit/
H A Djs-apis-mindSporeLite.md123 let modelBuffer = buffer.buffer;
163 let modelBuffer = buffer.buffer;
210 let modelBuffer = buffer.buffer;
365 resMgr.getRawFileContent(modelFile).then(modelBuffer => {
769 let modelBuffer = buffer.buffer;
774 modelInputs[0].setData(modelBuffer);
819 let modelBuffer = buffer.buffer;
823 modelInputs[0].setData(modelBuffer);
920 resMgr.getRawFileContent(modelFile).then((modelBuffer) => {
963 resMgr.getRawFileContent(modelFile).then((modelBuffer) => {
[all …]
/ohos5.0/foundation/ai/neural_network_runtime/test/fuzztest/
H A Dnnrt_device_impl.h90 …int32_t PrepareOfflineModel(const std::vector<SharedBuffer>& modelBuffer, const ModelConfig& confi… in PrepareOfflineModel() argument
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/include/
H A Dprepared_model_service.h42 int32_t Compile(const void* modelBuffer, size_t length);
/ohos5.0/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/include/
H A Dprepared_model_service.h42 NNRT_ReturnCode Compile(const void* modelBuffer, size_t length);
H A Dnnrt_device_service.h56 …int32_t PrepareOfflineModel(const std::vector<SharedBuffer>& modelBuffer, const ModelConfig& confi…
/ohos5.0/foundation/ai/neural_network_runtime/interfaces/kits/c/neural_network_runtime/
H A Dneural_network_core.h111 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_t …
/ohos5.0/foundation/ai/neural_network_runtime/frameworks/native/neural_network_core/
H A Dneural_network_core.cpp153 NNRT_API OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer in OH_NNCompilation_ConstructWithOfflineModelBuffer() argument
155 if (modelBuffer == nullptr) { in OH_NNCompilation_ConstructWithOfflineModelBuffer()
171 compilation->offlineModelBuffer.first = const_cast<void*>(modelBuffer); in OH_NNCompilation_ConstructWithOfflineModelBuffer()
/ohos5.0/docs/zh-cn/application-dev/reference/apis-neural-network-runtime-kit/
H A Dneural__network__core_8h.md30 …ime.md#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
H A D_neural_network_runtime.md81 …ffer](#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
1060 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer (const void *modelBuffer, size_t…
1069 …tion](#oh_nncompilation)实例只将**modelBuffer**指针保存在里面,而不是复制其数据。 在销毁[OH_NNCompilation](#oh_nncompilati…
1077 | modelBuffer | 离线模型文件内存。 |
H A D_neural_nework_runtime.md81 …ffer](#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
1060 OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer (const void *modelBuffer, size_t…
1069 …tion](#oh_nncompilation)实例只将**modelBuffer**指针保存在里面,而不是复制其数据。 在销毁[OH_NNCompilation](#oh_nncompilati…
1077 | modelBuffer | 离线模型文件内存。 |
/ohos5.0/docs/en/application-dev/reference/apis-neural-network-runtime-kit/
H A Dneural__network__core_8h.md30 …ime.md#oh_nncompilation_constructwithofflinemodelbuffer) (const void \*modelBuffer, size_t modelSi…
/ohos5.0/foundation/ai/neural_network_runtime/
H A Dneural-network-runtime-guidelines.md79 | OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_…
/ohos5.0/docs/zh-cn/application-dev/ai/nnrt/
H A Dneural-network-runtime-guidelines.md79 | OH_NNCompilation *OH_NNCompilation_ConstructWithOfflineModelBuffer(const void *modelBuffer, size_…

12