# bolt **Repository Path**: feng_wei_feng/bolt ## Basic Information - **Project Name**: bolt - **Description**: No description available - **Primary Language**: Unknown - **License**: MIT - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-01-06 - **Last Updated**: 2025-01-06 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Introduction --- [](https://opensource.org/licenses/MIT) [Bolt](https://huawei-noah.github.io/bolt/) is a light-weight library for deep learning. Bolt, as a universal deployment tool for all kinds of neural networks, aims to automate the deployment pipeline and achieve extreme acceleration. Bolt has been widely deployed and used in many departments of HUAWEI company, such as 2012 Laboratory, CBG and HUAWEI Product Lines. If you have questions or suggestions, you can submit issue. **QQ群: 833345709** # Why Bolt is what you need? --- - **High Performance:** **15%+** faster than existing open source acceleration libraries. - **Rich Model Conversion:** support Caffe, ONNX, TFLite, Tensorflow. - **Various Inference Precision:** support FP32, FP16, INT8, 1-BIT. - **Multiple platforms:** ARM CPU(v7, v8, v8.2+, v9), X86 CPU(AVX2, AVX512), GPU(Mali, Qualcomm, Intel, AMD) - **Bolt is the first to support NLP and also supports common CV applications.** - **Minimize ROM/RAM** - Rich Graph Optimization - Efficient Thread Affinity Setting - [Auto Algorithm Tuning](https://zhuanlan.zhihu.com/p/336218879) - [Time-Series Data Acceleration](docs/USER_HANDBOOK.md#time-series-data-acceleration) [See more excellent features and details here](https://zhuanlan.zhihu.com/p/317111024) # Building Status --- There are some common used platform for inference. More targets can be seen from [scripts/target.sh](scripts/target.sh). Please make a suitable choice depending on your environment. If you want to build on-device training module, you can add **--train** option. If you want to use multi-threads parallel, you can add **--openmp** option. If you want to build for cortex-M or cortex-A7 with restricted ROM/RAM(Sensor, MCU), you can see [docs/LITE.md](docs/LITE.md). *Bolt defaultly link static library, This may cause some problem on some platforms. You can use --shared option to link shared library.* | target platform | precision | build command | Linux | Windows | MacOS | | ---------------------- | ------------------ | ---------------------------------------------------- | ----- | ------- | ----- | | Android(armv7) | fp32,int8 | ./install.sh --target=android-armv7 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-armv7) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-armv7) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-armv7) | | Android(armv8) | fp32,int8 | ./install.sh --target=android-aarch64 --fp16=off | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-armv8) | | Android(armv8.2+) | fp32,fp16,int8,bnn | ./install.sh --target=android-aarch64 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-armv8) | | Android(armv9) | fp32,fp16,bf16,int8,bnn | ./install.sh --target=android-aarch64_v9 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-armv8) | | Android(gpu) | fp16 | ./install.sh --target=android-aarch64 --gpu | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-armv8) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-armv8) | | Android(x86_64) | fp32,int8 | ./install.sh --target=android-x86_64 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-android-x86_64) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-android-x86_64) | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-android-x86_64) | | iOS(armv7) | fp32,int8 | ./install.sh --target=ios-armv7 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-ios-armv7) | | iOS(armv8) | fp32,int8 | ./install.sh --target=ios-aarch64 --fp16=off | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-ios-armv8) | | iOS(armv8.2+) | fp32,fp16,int8,bnn | ./install.sh --target=ios-aarch64 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-ios-armv8) | | Linux(armv7) | fp32,int8 | ./install.sh --target=linux-armv7_blank | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86) | / | / | | Linux(armv8) | fp32,int8 | ./install.sh --target=linux-aarch64_blank --fp16=off | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86) | / | / | | Linux(armv8.2+) | fp32,fp16,int8,bnn | ./install.sh --target=linux-aarch64_blank | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86) | / | / | | Linux(x86_64) | fp32,int8 | ./install.sh --target=linux-x86_64 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86) | / | / | | Linux(x86_64_avx2) | fp32 | ./install.sh --target=linux-x86_64_avx2 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86-avx2) | / | / | | Linux(x86_64_avx512) | fp32,int8 | ./install.sh --target=linux-x86_64_avx512 | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Alinux-x86-avx2) | / | / | | Windows(x86_64) | fp32,int8 | ./install.sh --target=windows-x86_64 | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-x86) | / | | Windows(x86_64_avx2) | fp32 | ./install.sh --target=windows-x86_64_avx2 | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-x86-avx2) | / | | Windows(gpu) | fp16 | ./install.sh --target=windows-x86_64_avx2 --gpu --fp16=on | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-x86-avx2) | / | | Windows(x86_64_avx512) | fp32,int8 | ./install.sh --target=windows-x86_64_avx512 | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Awindows-x86-avx2) | / | | Windows(armv8.2+) | fp32,fp16,int8,bnn | ./install.sh --target=windows-aarch64 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-x86) | | MacOS(x86_64) | fp32,int8 | ./install.sh --target=macos-x86_64 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-x86) | | MacOS(x86_64_avx2) | fp32 | ./install.sh --target=macos-x86_64_avx2 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-x86-avx2) | | MacOS(x86_64_avx512) | fp32,int8 | ./install.sh --target=macos-x86_64_avx512 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-x86-avx2) | | MacOS(armv8.2+) | fp32,fp16,int8,bnn | ./install.sh --target=macos-aarch64 | / | / | [](https://github.com/huawei-noah/bolt/actions?query=workflow%3Amacos-x86) | # Quick Start ---

|
|
|
|
|
|
|
# Verified Networks
---
Bolt has shown its high performance in the inference of common CV, NLP and Recommendation neural networks. Some of the representative networks that we have verified are listed below. You can find detailed benchmark information in [docs/BENCHMARK.md](docs/BENCHMARK.md).
| Application | Models |
| CV | Resnet50, Shufflenet, Squeezenet, Densenet, Efficientnet, Mobilenet_v1, Mobilenet_v2, Mobilenet_v3, BiRealNet, ReActNet, Ghostnet, unet, LCNet, Pointnet, hair-segmentation, duc, fcn, retinanet, SSD, Faster-RCNN, Mask-RCNN, Yolov2, Yolov3, Yolov4, Yolov5, ViT, TNT, RepVGG, VitAE, CMT, EfficientFormer ... |
| NLP | Bert, Albert, Tinybert, Neural Machine Translation, Text To Speech(Tactron,Tactron2,FastSpeech+hifigan,melgan), Automatic Speech Recognition, DFSMN, Conformer, Tdnn, FRILL, T5, GPT-2, Roberta, Wenet ... |
| Recommendation | NFM, AFM, ONN, wide&deep, DeepFM, MMOE |
| More DL Tasks | ... |