diff --git a/frameworks/native/neural_network_runtime/neural_network_runtime.cpp b/frameworks/native/neural_network_runtime/neural_network_runtime.cpp index 39f79d236c32e024918f511e310bded47169dfa2..cb97433af6f8ec561a5d01a2240c304ed76a5367 100644 --- a/frameworks/native/neural_network_runtime/neural_network_runtime.cpp +++ b/frameworks/native/neural_network_runtime/neural_network_runtime.cpp @@ -566,7 +566,8 @@ NNRT_API OH_NN_ReturnCode OH_NNModel_BuildFromLiteGraph(OH_NNModel *model, const } namespace { -OH_NN_ReturnCode CheckCacheFileExtension(const std::string& content, int64_t& fileNumber, int64_t& cacheVersion, int64_t& deviceId) +OH_NN_ReturnCode CheckCacheFileExtension(const std::string& content, int64_t& fileNumber, + int64_t& cacheVersion, int64_t& deviceId) { if (!nlohmann::json::accept(content)) { LOGE("OH_NNModel_HasCache CheckCacheFile JSON parse error"); @@ -617,7 +618,8 @@ OH_NN_ReturnCode CheckCacheFileExtension(const std::string& content, int64_t& fi return OH_NN_SUCCESS; } -OH_NN_ReturnCode CheckCacheFile(const std::string& cacheInfoPath, int64_t& fileNumber, int64_t& cacheVersion, int64_t& deviceId) +OH_NN_ReturnCode CheckCacheFile(const std::string& cacheInfoPath, int64_t& fileNumber, + int64_t& cacheVersion, int64_t& deviceId) { char path[PATH_MAX]; if (realpath(cacheInfoPath.c_str(), path) == nullptr) {