diff --git a/README.md b/README.md index 499ace5d392d55ce7891a09bcc76f980d55acd52..f421dfd0772671b9ef4da8d840f8956b52dedb7a 100644 --- a/README.md +++ b/README.md @@ -1,5 +1,15 @@ +[](README_en.md) [](README.md) + # DeepSparkInference +
+ Homepage + + +
+
+ DeepSparkInference推理模型库作为DeepSpark开源社区的核心项目,于2024年3月正式开源,一期甄选了48个推理模型示例,涵盖计算机视觉,自然语言处理,语音识别等领域,后续将逐步拓展更多AI领域。 DeepSparkInference中的模型提供了在国产推理引擎IGIE或IxRT下运行的推理示例和指导文档,部分模型提供了基于国产通用GPU[智铠100](https://www.iluvatar.com/productDetails?fullCode=cpjs-yj-tlxltt-zk100)的评测结果。 @@ -10,7 +20,9 @@ IxRT(Iluvatar CoreX RunTime)是天数智芯自研的高性能推理引擎, DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型类别并拓展大模型推理。 -## LLM (Large Language Model) +## 模型库 + +### 大语言模型(LLM) | Model | vLLM | TRT-LLM | TGI | IXUCA SDK | |-------------------------------|--------------------------------------------------------|---------------------------------------|------------------------------------|-----------| @@ -36,9 +48,9 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | Qwen2-72B Instruct | [✅](models/nlp/llm/qwen2-72b/vllm) | | | 4.2.0 | | StableLM2-1.6B | [✅](models/nlp/llm/stablelm/vllm) | | | 4.2.0 | -## Computer Vision +### 计算机视觉(CV) -### Classification +#### 视觉分类 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |------------------------|-------|--------------------------------------------------------|-----------------------------------------------------------|-----------| @@ -127,7 +139,7 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | | INT8 | [✅](models/cv/classification/wide_resnet50/igie) | [✅](models/cv/classification/wide_resnet50/ixrt) | 4.2.0 | | Wide ResNet101 | FP16 | [✅](models/cv/classification/wide_resnet101/igie) | | 4.2.0 | -### Object Detection +#### 目标检测 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |------------|-------|-------------------------------------------------|-------------------------------------------------|-----------| @@ -163,21 +175,21 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | YOLOX | FP16 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.2.0 | | | INT8 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.2.0 | -### Face Recognition +#### 人脸识别 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |---------|-------|------|----------------------------------------------|-----------| | FaceNet | FP16 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.2.0 | | | INT8 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.2.0 | -### OCR (Optical Character Recognition) +#### 光学字符识别(OCR) | Model | Prec. | IGIE | IXUCA SDK | |---------------|-------|---------------------------------------|-----------| | Kie_layoutXLM | FP16 | [✅](models/cv/ocr/kie_layoutxlm/igie) | 4.2.0 | | SVTR | FP16 | [✅](models/cv/ocr/svtr/igie) | 4.2.0 | -### Pose Estimation +#### 姿态估计 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |----------------------|-------|-----------------------------------------------|----------------------------------------------------------|-----------| @@ -185,14 +197,14 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | Lightweight OpenPose | FP16 | | [✅](models/cv/pose_estimation/lightweight_openpose/ixrt) | 4.2.0 | | RTMPose | FP16 | [✅](models/cv/pose_estimation/rtmpose/igie) | [✅](models/cv/pose_estimation/rtmpose/ixrt) | 4.2.0 | -### Instance Segmentation +#### 实例分割 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |------------|-------|------|-----------------------------------------------------|-----------| | Mask R-CNN | FP16 | | [✅](models/cv/instance_segmentation/mask_rcnn/ixrt) | 4.2.0 | | SOLOv1 | FP16 | | [✅](models/cv/instance_segmentation/solov1/ixrt) | 4.2.0 | -### Multi-Object Tracking +#### 多目标跟踪 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |---------------------|-------|----------------------------------------------------|------|-----------| @@ -201,7 +213,7 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | | INT8 | [✅](models/cv/multi_object_tracking/deepsort/igie) | | 4.2.0 | | RepNet-Vehicle-ReID | FP16 | [✅](models/cv/multi_object_tracking/repnet/igie) | | 4.2.0 | -## Multimodal +### 多模态 | Model | vLLM | IxFormer | IXUCA SDK | |---------------------|-----------------------------------------------------------------------|------------------------------------------------------------|-----------| @@ -213,9 +225,9 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | LLaVA-Next-Video-7B | [✅](models/multimodal/vision_language_model/llava_next_video_7b/vllm) | | 4.2.0 | | MiniCPM V2 | [✅](models/multimodal/vision_language_model/minicpm_v_2/vllm) | | 4.2.0 | -## NLP +### 自然语言处理(NLP) -### PLM (Pre-trained Language Model) +#### 预训练语言模型(PLM) | Model | Prec. | IGIE | IxRT | IXUCA SDK | |------------------|-------|-------------------------------------------|-------------------------------------------|-----------| @@ -230,18 +242,18 @@ DeepSparkInference将按季度进行版本更新,后续会逐步丰富模型 | RoFormer | FP16 | | [✅](models/nlp/plm/roformer/ixrt) | 4.2.0 | | VideoBERT | FP16 | | [✅](models/nlp/plm/videobert/ixrt) | 4.2.0 | -## Audio +### 语音 -### Speech Recognition +#### 语音识别 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |-----------------|-------|-----------------------------------------------------|-----------------------------------------------------------|-----------| | Conformer | FP16 | [✅](models/audio/speech_recognition/conformer/igie) | [✅](models/audio/speech_recognition/conformer/ixrt) | 4.2.0 | | Transformer ASR | FP16 | | [✅](models/audio/speech_recognition/transformer_asr/ixrt) | 4.2.0 | -## Others +### 其他 -### Recommendation Systems +#### 推荐系统 | Model | Prec. | IGIE | IxRT | IXUCA SDK | |-------------|-------|------|------------------------------------------------------|-----------| diff --git a/README_en.md b/README_en.md new file mode 100644 index 0000000000000000000000000000000000000000..df4d7aba4d8be107bb869dc6f66d622231e6af04 --- /dev/null +++ b/README_en.md @@ -0,0 +1,305 @@ +[](README_en.md) [](README.md) + +# DeepSparkInference + +
+ Homepage + + +
+
+ +DeepSparkInference ModelZoo, as a core project of the DeepSpark open-source community, was officially open-sourced in +March 2024. The first release selected 48 inference model examples, covering fields such as computer vision, natural +language processing, and speech recognition. More AI domains will be gradually expanded in the future. + +The models in DeepSparkInference provide inference examples and guidance documents for running on inference engines IGIE +or IxRT self-developed by Iluvatar CoreX. Some models provide evaluation results based on the self-developed GPGPU +Zhikai 100. + +IGIE (Iluvatar GPU Inference Engine) is a high-performance, highly gene, and end-to-end AI inference engine developed +based on the TVM framework. It supports multi-framework model, quantization, graph optimization, multi-operator library +support, multi-backend support, and automatic operator tuning, providing an easy-to-deploy, high-throughput, and +low-latency complete solution for inference scenarios. + +IxRT (Iluvatar CoreX RunTime) is a high-performance inference engine independently developed by Iluvatar CoreX, focusing +on maximizing the performance of Iluvatar CoreX's GPGPU and achieving high-performance inference for models in various +fields. IxRT supports features such as dynamic shape inference, plugins, and INT8/FP16 inference. + +DeepSparkInference will be updated quarterly, and model categories will be gradually enriched, with large model +inference to be expanded in the future. + +## ModelZoo + +### LLM (Large Language Model) + +| Model | vLLM | TRT-LLM | TGI | IXUCA SDK | +|-------------------------------|--------------------------------------------------------|---------------------------------------|------------------------------------|-----------| +| Baichuan2-7B | [✅](models/nlp/llm/baichuan2-7b/vllm) | | | 4.2.0 | +| ChatGLM-3-6B | [✅](models/nlp/llm/chatglm3-6b/vllm) | | | 4.2.0 | +| ChatGLM-3-6B-32K | [✅](models/nlp/llm/chatglm3-6b-32k/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Llama-8B | [✅](models/nlp/llm/deepseek-r1-distill-llama-8b/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Llama-70B | [✅](models/nlp/llm/deepseek-r1-distill-llama-70b/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Qwen-1.5B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-1.5b/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Qwen-7B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-7b/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Qwen-14B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-14b/vllm) | | | 4.2.0 | +| DeepSeek-R1-Distill-Qwen-32B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-32b/vllm) | | | 4.2.0 | +| Llama2-7B | [✅](models/nlp/llm/llama2-7b/vllm) | [✅](models/nlp/llm/llama2-7b/trtllm) | | 4.2.0 | +| Llama2-13B | | [✅](models/nlp/llm/llama2-13b/trtllm) | | 4.2.0 | +| Llama2-70B | | [✅](models/nlp/llm/llama2-70b/trtllm) | | 4.2.0 | +| Llama3-70B | [✅](models/nlp/llm/llama3-70b/vllm) | | | 4.2.0 | +| Qwen-7B | [✅](models/nlp/llm/qwen-7b/vllm) | | | 4.2.0 | +| Qwen1.5-7B | [✅](models/nlp/llm/qwen1.5-7b/vllm) | | [✅](models/nlp/llm/qwen1.5-7b/tgi) | 4.2.0 | +| Qwen1.5-14B | [✅](models/nlp/llm/qwen1.5-14b/vllm) | | | 4.2.0 | +| Qwen1.5-32B Chat | [✅](models/nlp/llm/qwen1.5-32b/vllm) | | | 4.2.0 | +| Qwen1.5-72B | [✅](models/nlp/llm/qwen1.5-72b/vllm) | | | 4.2.0 | +| Qwen2-7B Instruct | [✅](models/nlp/llm/qwen2-7b/vllm) | | | 4.2.0 | +| Qwen2-72B Instruct | [✅](models/nlp/llm/qwen2-72b/vllm) | | | 4.2.0 | +| StableLM2-1.6B | [✅](models/nlp/llm/stablelm/vllm) | | | 4.2.0 | + +### Computer Vision + +#### Classification + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|------------------------|-------|--------------------------------------------------------|-----------------------------------------------------------|-----------| +| AlexNet | FP16 | [✅](models/cv/classification/alexnet/igie) | [✅](models/cv/classification/alexnet/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/alexnet/igie) | [✅](models/cv/classification/alexnet/ixrt) | 4.2.0 | +| CLIP | FP16 | [✅](models/cv/classification/clip/igie) | | 4.2.0 | +| Conformer-B | FP16 | [✅](models/cv/classification/conformer_base/igie) | | 4.2.0 | +| ConvNeXt-Base | FP16 | [✅](models/cv/classification/convnext_base/igie) | [✅](models/cv/classification/convnext_base/ixrt) | 4.2.0 | +| ConvNext-S | FP16 | [✅](models/cv/classification/convnext_s/igie) | | 4.2.0 | +| ConvNeXt-Small | FP16 | [✅](models/cv/classification/convnext_small/igie) | [✅](models/cv/classification/convnext_small/ixrt) | 4.2.0 | +| CSPDarkNet53 | FP16 | [✅](models/cv/classification/cspdarknet53/igie) | [✅](models/cv/classification/cspdarknet53/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/cspdarknet53/ixrt) | 4.2.0 | +| CSPResNet50 | FP16 | [✅](models/cv/classification/cspresnet50/igie) | [✅](models/cv/classification/cspresnet50/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/cspresnet50/ixrt) | 4.2.0 | +| DeiT-tiny | FP16 | [✅](models/cv/classification/deit_tiny/igie) | [✅](models/cv/classification/deit_tiny/ixrt) | 4.2.0 | +| DenseNet121 | FP16 | [✅](models/cv/classification/densenet121/igie) | [✅](models/cv/classification/densenet121/ixrt) | 4.2.0 | +| DenseNet161 | FP16 | [✅](models/cv/classification/densenet161/igie) | [✅](models/cv/classification/densenet161/ixrt) | 4.2.0 | +| DenseNet169 | FP16 | [✅](models/cv/classification/densenet169/igie) | [✅](models/cv/classification/densenet169/ixrt) | 4.2.0 | +| DenseNet201 | FP16 | [✅](models/cv/classification/densenet201/igie) | [✅](models/cv/classification/densenet201/ixrt) | 4.2.0 | +| EfficientNet-B0 | FP16 | [✅](models/cv/classification/efficientnet_b0/igie) | [✅](models/cv/classification/efficientnet_b0/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/efficientnet_b0/ixrt) | 4.2.0 | +| EfficientNet-B1 | FP16 | [✅](models/cv/classification/efficientnet_b1/igie) | [✅](models/cv/classification/efficientnet_b1/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/efficientnet_b1/ixrt) | 4.2.0 | +| EfficientNet-B2 | FP16 | [✅](models/cv/classification/efficientnet_b2/igie) | [✅](models/cv/classification/efficientnet_b2/ixrt) | 4.2.0 | +| EfficientNet-B3 | FP16 | [✅](models/cv/classification/efficientnet_b3/igie) | [✅](models/cv/classification/efficientnet_b3/ixrt) | 4.2.0 | +| EfficientNet-B4 | FP16 | [✅](models/cv/classification/efficientnet_b4/igie) | | 4.2.0 | +| EfficientNetV2 | FP16 | [✅](models/cv/classification/efficientnet_v2/igie) | [✅](models/cv/classification/efficientnet_v2/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/efficientnet_v2/ixrt) | 4.2.0 | +| EfficientNetv2_rw_t | FP16 | [✅](models/cv/classification/efficientnetv2_rw_t/igie) | [✅](models/cv/classification/efficientnetv2_rw_t/ixrt) | 4.2.0 | +| EfficientNetv2_s | FP16 | [✅](models/cv/classification/efficientnet_v2_s/igie) | [✅](models/cv/classification/efficientnet_v2_s/ixrt) | 4.2.0 | +| GoogLeNet | FP16 | [✅](models/cv/classification/googlenet/igie) | [✅](models/cv/classification/googlenet/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/googlenet/igie) | [✅](models/cv/classification/googlenet/ixrt) | 4.2.0 | +| HRNet-W18 | FP16 | [✅](models/cv/classification/hrnet_w18/igie) | [✅](models/cv/classification/hrnet_w18/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/hrnet_w18/ixrt) | 4.2.0 | +| InceptionV3 | FP16 | [✅](models/cv/classification/inception_v3/igie) | [✅](models/cv/classification/inception_v3/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/inception_v3/igie) | [✅](models/cv/classification/inception_v3/ixrt) | 4.2.0 | +| Inception-ResNet-V2 | FP16 | | [✅](models/cv/classification/inception_resnet_v2/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/inception_resnet_v2/ixrt) | 4.2.0 | +| Mixer_B | FP16 | [✅](models/cv/classification/mlp_mixer_base/igie) | | 4.2.0 | +| MNASNet0_5 | FP16 | [✅](models/cv/classification/mnasnet0_5/igie) | | 4.2.0 | +| MNASNet0_75 | FP16 | [✅](models/cv/classification/mnasnet0_75/igie) | | 4.2.0 | +| MobileNetV2 | FP16 | [✅](models/cv/classification/mobilenet_v2/igie) | [✅](models/cv/classification/mobilenet_v2/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/mobilenet_v2/igie) | [✅](models/cv/classification/mobilenet_v2/ixrt) | 4.2.0 | +| MobileNetV3_Large | FP16 | [✅](models/cv/classification/mobilenet_v3_large/igie) | | 4.2.0 | +| MobileNetV3_Small | FP16 | [✅](models/cv/classification/mobilenet_v3/igie) | [✅](models/cv/classification/mobilenet_v3/ixrt) | 4.2.0 | +| MViTv2_base | FP16 | [✅](models/cv/classification/mvitv2_base/igie) | | 4.2.0 | +| RegNet_x_16gf | FP16 | [✅](models/cv/classification/regnet_x_16gf/igie) | | 4.2.0 | +| RegNet_x_1_6gf | FP16 | [✅](models/cv/classification/regnet_x_1_6gf/igie) | | 4.2.0 | +| RegNet_y_1_6gf | FP16 | [✅](models/cv/classification/regnet_y_1_6gf/igie) | | 4.2.0 | +| RepVGG | FP16 | [✅](models/cv/classification/repvgg/igie) | [✅](models/cv/classification/repvgg/ixrt) | 4.2.0 | +| Res2Net50 | FP16 | [✅](models/cv/classification/res2net50/igie) | [✅](models/cv/classification/res2net50/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/res2net50/ixrt) | 4.2.0 | +| ResNeSt50 | FP16 | [✅](models/cv/classification/resnest50/igie) | | 4.2.0 | +| ResNet101 | FP16 | [✅](models/cv/classification/resnet101/igie) | [✅](models/cv/classification/resnet101/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/resnet101/igie) | [✅](models/cv/classification/resnet101/ixrt) | 4.2.0 | +| ResNet152 | FP16 | [✅](models/cv/classification/resnet152/igie) | | 4.2.0 | +| | INT8 | [✅](models/cv/classification/resnet152/igie) | | 4.2.0 | +| ResNet18 | FP16 | [✅](models/cv/classification/resnet18/igie) | [✅](models/cv/classification/resnet18/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/resnet18/igie) | [✅](models/cv/classification/resnet18/ixrt) | 4.2.0 | +| ResNet34 | FP16 | | [✅](models/cv/classification/resnet34/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/resnet34/ixrt) | 4.2.0 | +| ResNet50 | FP16 | [✅](models/cv/classification/resnet50/igie) | [✅](models/cv/classification/resnet50/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/resnet50/igie) | | 4.2.0 | +| ResNetV1D50 | FP16 | [✅](models/cv/classification/resnetv1d50/igie) | [✅](models/cv/classification/resnetv1d50/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/resnetv1d50/ixrt) | 4.2.0 | +| ResNeXt50_32x4d | FP16 | [✅](models/cv/classification/resnext50_32x4d/igie) | [✅](models/cv/classification/resnext50_32x4d/ixrt) | 4.2.0 | +| ResNeXt101_64x4d | FP16 | [✅](models/cv/classification/resnext101_64x4d/igie) | | 4.2.0 | +| ResNeXt101_32x8d | FP16 | [✅](models/cv/classification/resnext101_32x8d/igie) | | 4.2.0 | +| SEResNet50 | FP16 | [✅](models/cv/classification/se_resnet50/igie) | | 4.2.0 | +| ShuffleNetV1 | FP16 | | [✅](models/cv/classification/shufflenet_v1/ixrt) | 4.2.0 | +| ShuffleNetV2_x0_5 | FP16 | [✅](models/cv/classification/shufflenetv2_x0_5/igie) | | 4.2.0 | +| ShuffleNetV2_x1_0 | FP16 | [✅](models/cv/classification/shufflenetv2_x1_0/igie) | | 4.2.0 | +| ShuffleNetV2_x1_5 | FP16 | [✅](models/cv/classification/shufflenetv2_x1_5/igie) | | 4.2.0 | +| ShuffleNetV2_x2_0 | FP16 | [✅](models/cv/classification/shufflenetv2_x2_0/igie) | | 4.2.0 | +| SqueezeNet 1.0 | FP16 | [✅](models/cv/classification/squeezenet_v1_0/igie) | [✅](models/cv/classification/squeezenet_v1_0/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/squeezenet_v1_0/ixrt) | 4.2.0 | +| SqueezeNet 1.1 | FP16 | | [✅](models/cv/classification/squeezenet_v1_1/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/classification/squeezenet_v1_1/ixrt) | 4.2.0 | +| SVT Base | FP16 | [✅](models/cv/classification/svt_base/igie) | | 4.2.0 | +| Swin Transformer | FP16 | [✅](models/cv/classification/swin_transformer/igie) | | 4.2.0 | +| Swin Transformer Large | FP16 | | [✅](models/cv/classification/swin_transformer_large/ixrt) | 4.2.0 | +| VGG11 | FP16 | [✅](models/cv/classification/vgg11/igie) | | 4.2.0 | +| VGG16 | FP16 | [✅](models/cv/classification/vgg16/igie) | [✅](models/cv/classification/vgg16/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/vgg16/igie) | | 4.2.0 | +| Wide ResNet50 | FP16 | [✅](models/cv/classification/wide_resnet50/igie) | [✅](models/cv/classification/wide_resnet50/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/classification/wide_resnet50/igie) | [✅](models/cv/classification/wide_resnet50/ixrt) | 4.2.0 | +| Wide ResNet101 | FP16 | [✅](models/cv/classification/wide_resnet101/igie) | | 4.2.0 | + +#### Object Detection + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|------------|-------|-------------------------------------------------|-------------------------------------------------|-----------| +| ATSS | FP16 | [✅](models/cv/object_detection/atss/igie) | | 4.2.0 | +| CenterNet | FP16 | [✅](models/cv/object_detection/centernet/igie) | [✅](models/cv/object_detection/centernet/ixrt) | 4.2.0 | +| DETR | FP16 | | [✅](models/cv/object_detection/detr/ixrt) | 4.2.0 | +| FCOS | FP16 | [✅](models/cv/object_detection/fcos/igie) | [✅](models/cv/object_detection/fcos/ixrt) | 4.2.0 | +| FoveaBox | FP16 | [✅](models/cv/object_detection/foveabox/igie) | [✅](models/cv/object_detection/foveabox/ixrt) | 4.2.0 | +| FSAF | FP16 | [✅](models/cv/object_detection/fsaf/igie) | [✅](models/cv/object_detection/fsaf/ixrt) | 4.2.0 | +| HRNet | FP16 | [✅](models/cv/object_detection/hrnet/igie) | [✅](models/cv/object_detection/hrnet/ixrt) | 4.2.0 | +| PAA | FP16 | [✅](models/cv/object_detection/paa/igie) | | 4.2.0 | +| RetinaFace | FP16 | [✅](models/cv/object_detection/retinaface/igie) | [✅](models/cv/object_detection/retinaface/ixrt) | 4.2.0 | +| RetinaNet | FP16 | [✅](models/cv/object_detection/retinanet/igie) | | 4.2.0 | +| RTMDet | FP16 | [✅](models/cv/object_detection/rtmdet/igie) | | 4.2.0 | +| SABL | FP16 | [✅](models/cv/object_detection/sabl/igie) | | 4.2.0 | +| YOLOv3 | FP16 | [✅](models/cv/object_detection/yolov3/igie) | [✅](models/cv/object_detection/yolov3/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolov3/igie) | [✅](models/cv/object_detection/yolov3/ixrt) | 4.2.0 | +| YOLOv4 | FP16 | [✅](models/cv/object_detection/yolov4/igie) | [✅](models/cv/object_detection/yolov4/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolov4/igie16) | [✅](models/cv/object_detection/yolov4/ixrt16) | 4.2.0 | +| YOLOv5 | FP16 | [✅](models/cv/object_detection/yolov5/igie) | [✅](models/cv/object_detection/yolov5/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolov5/igie) | [✅](models/cv/object_detection/yolov5/ixrt) | 4.2.0 | +| YOLOv5s | FP16 | | [✅](models/cv/object_detection/yolov5s/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/object_detection/yolov5s/ixrt) | 4.2.0 | +| YOLOv6 | FP16 | [✅](models/cv/object_detection/yolov6/igie) | [✅](models/cv/object_detection/yolov6/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/object_detection/yolov6/ixrt) | 4.2.0 | +| YOLOv7 | FP16 | [✅](models/cv/object_detection/yolov7/igie) | [✅](models/cv/object_detection/yolov7/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolov7/igie) | [✅](models/cv/object_detection/yolov7/ixrt) | 4.2.0 | +| YOLOv8 | FP16 | [✅](models/cv/object_detection/yolov8/igie) | [✅](models/cv/object_detection/yolov8/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolov8/igie) | [✅](models/cv/object_detection/yolov8/ixrt) | 4.2.0 | +| YOLOv9 | FP16 | [✅](models/cv/object_detection/yolov9/igie) | | 4.2.0 | +| YOLOv10 | FP16 | [✅](models/cv/object_detection/yolov10/igie) | | 4.2.0 | +| YOLOv11 | FP16 | [✅](models/cv/object_detection/yolov11/igie) | | 4.2.0 | +| YOLOX | FP16 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.2.0 | +| | INT8 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.2.0 | + +#### Face Recognition + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|---------|-------|------|----------------------------------------------|-----------| +| FaceNet | FP16 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.2.0 | + +#### OCR (Optical Character Recognition) + +| Model | Prec. | IGIE | IXUCA SDK | +|---------------|-------|---------------------------------------|-----------| +| Kie_layoutXLM | FP16 | [✅](models/cv/ocr/kie_layoutxlm/igie) | 4.2.0 | +| SVTR | FP16 | [✅](models/cv/ocr/svtr/igie) | 4.2.0 | + +#### Pose Estimation + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|----------------------|-------|-----------------------------------------------|----------------------------------------------------------|-----------| +| HRNetPose | FP16 | [✅](models/cv/pose_estimation/hrnetpose/igie) | | 4.2.0 | +| Lightweight OpenPose | FP16 | | [✅](models/cv/pose_estimation/lightweight_openpose/ixrt) | 4.2.0 | +| RTMPose | FP16 | [✅](models/cv/pose_estimation/rtmpose/igie) | [✅](models/cv/pose_estimation/rtmpose/ixrt) | 4.2.0 | + +#### Instance Segmentation + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|------------|-------|------|-----------------------------------------------------|-----------| +| Mask R-CNN | FP16 | | [✅](models/cv/instance_segmentation/mask_rcnn/ixrt) | 4.2.0 | +| SOLOv1 | FP16 | | [✅](models/cv/instance_segmentation/solov1/ixrt) | 4.2.0 | + +#### Multi-Object Tracking + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|---------------------|-------|----------------------------------------------------|------|-----------| +| FastReID | FP16 | [✅](models/cv/multi_object_tracking/fastreid/igie) | | 4.2.0 | +| DeepSort | FP16 | [✅](models/cv/multi_object_tracking/deepsort/igie) | | 4.2.0 | +| | INT8 | [✅](models/cv/multi_object_tracking/deepsort/igie) | | 4.2.0 | +| RepNet-Vehicle-ReID | FP16 | [✅](models/cv/multi_object_tracking/repnet/igie) | | 4.2.0 | + +### Multimodal + +| Model | vLLM | IxFormer | IXUCA SDK | +|---------------------|-----------------------------------------------------------------------|------------------------------------------------------------|-----------| +| Chameleon-7B | [✅](models/multimodal/vision_language_model/chameleon_7b/vllm) | | 4.2.0 | +| CLIP | | [✅](models/multimodal/vision_language_model/clip/ixformer) | 4.2.0 | +| Fuyu-8B | [✅](models/multimodal/vision_language_model/fuyu_8b/vllm) | | 4.2.0 | +| InternVL2-4B | [✅](models/multimodal/vision_language_model/intern_vl/vllm) | | 4.2.0 | +| LLaVA | [✅](models/multimodal/vision_language_model/llava/vllm) | | 4.2.0 | +| LLaVA-Next-Video-7B | [✅](models/multimodal/vision_language_model/llava_next_video_7b/vllm) | | 4.2.0 | +| MiniCPM V2 | [✅](models/multimodal/vision_language_model/minicpm_v_2/vllm) | | 4.2.0 | + +### NLP + +#### PLM (Pre-trained Language Model) + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|------------------|-------|-------------------------------------------|-------------------------------------------|-----------| +| ALBERT | FP16 | | [✅](models/nlp/plm/albert/ixrt) | 4.2.0 | +| BERT Base NER | INT8 | [✅](models/nlp/plm/bert_base_ner/igie) | | 4.2.0 | +| BERT Base SQuAD | FP16 | [✅](models/nlp/plm/bert_base_squad/igie) | [✅](models/nlp/plm/bert_base_squad/ixrt) | 4.2.0 | +| | INT8 | | [✅](models/nlp/plm/bert_base_squad/ixrt) | 4.2.0 | +| BERT Large SQuAD | FP16 | [✅](models/nlp/plm/bert_large_squad/igie) | [✅](models/nlp/plm/bert_large_squad/ixrt) | 4.2.0 | +| | INT8 | [✅](models/nlp/plm/bert_large_squad/igie) | [✅](models/nlp/plm/bert_large_squad/ixrt) | 4.2.0 | +| DeBERTa | FP16 | | [✅](models/nlp/plm/deberta/ixrt) | 4.2.0 | +| RoBERTa | FP16 | | [✅](models/nlp/plm/roberta/ixrt) | 4.2.0 | +| RoFormer | FP16 | | [✅](models/nlp/plm/roformer/ixrt) | 4.2.0 | +| VideoBERT | FP16 | | [✅](models/nlp/plm/videobert/ixrt) | 4.2.0 | + +### Audio + +#### Speech Recognition + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|-----------------|-------|-----------------------------------------------------|-----------------------------------------------------------|-----------| +| Conformer | FP16 | [✅](models/audio/speech_recognition/conformer/igie) | [✅](models/audio/speech_recognition/conformer/ixrt) | 4.2.0 | +| Transformer ASR | FP16 | | [✅](models/audio/speech_recognition/transformer_asr/ixrt) | 4.2.0 | + +### Others + +#### Recommendation Systems + +| Model | Prec. | IGIE | IxRT | IXUCA SDK | +|-------------|-------|------|------------------------------------------------------|-----------| +| Wide & Deep | FP16 | | [✅](models/others/recommendation/wide_and_deep/ixrt) | 4.2.0 | + +--- + +## Community + +### Code of Conduct + +Please refer to DeepSpark Code of Conduct on +[Gitee](https://gitee.com/deep-spark/deepspark/blob/master/CODE_OF_CONDUCT.md) or on +[GitHub](https://github.com/Deep-Spark/deepspark/blob/main/CODE_OF_CONDUCT.md). + +### Contact + +Please contact . + +### Contribution + +Please refer to the [DeepSparkInference Contributing Guidelines](CONTRIBUTING.md). + +### Disclaimers + +DeepSparkInference only provides download and preprocessing scripts for public datasets. These datasets do not belong to +DeepSparkInference, and DeepSparkInference is not responsible for their quality or maintenance. Please ensure that you +have the necessary usage licenses for these datasets. Models trained based on these datasets can only be used for +non-commercial research and education purposes. + +To dataset owners: + +If you do not want your dataset to be published on DeepSparkInference or wish to update the dataset that belongs to you +on DeepSparkInference, please submit an issue on Gitee or Github. We will delete or update it according to your issue. +We sincerely appreciate your support and contributions to our community. + +## License + +This project is released under [Apache-2.0](LICENSE) License. diff --git a/models/cv/classification/efficientnet_v2/ixrt/README.md b/models/cv/classification/efficientnet_v2/ixrt/README.md index 9c85fb70b236d19e4bb69d21f721d9badd9c5143..1e131e4dd81b1508ccb5e48958997dd33f375685 100755 --- a/models/cv/classification/efficientnet_v2/ixrt/README.md +++ b/models/cv/classification/efficientnet_v2/ixrt/README.md @@ -36,10 +36,9 @@ pip3 install -r requirements.txt ```bash mkdir checkpoints -git clone https://github.com/huggingface/pytorch-image-models.git +git clone -b v1.0.11 https://github.com/huggingface/pytorch-image-models.git cp ./export_onnx.py pytorch-image-models/timm/models -rm pytorch-image-models/timm/models/_builder.py -mv ./_builder.py pytorch-image-models/timm/models +cp ./_builder.py pytorch-image-models/timm/models cd pytorch-image-models/timm mkdir -p /root/.cache/torch/hub/checkpoints/ wget -P /root/.cache/torch/hub/checkpoints/ https://github.com/rwightman/pytorch-image-models/releases/download/v0.1-weights/efficientnetv2_t_agc-3620981a.pth diff --git a/models/cv/classification/efficientnet_v2/ixrt/_builder.py b/models/cv/classification/efficientnet_v2/ixrt/_builder.py index 7246c0d57abaffac744e861b959b9169121265c3..9e8c817afcf49e4d3633e178b4b840aee62d2f8d 100755 --- a/models/cv/classification/efficientnet_v2/ixrt/_builder.py +++ b/models/cv/classification/efficientnet_v2/ixrt/_builder.py @@ -191,7 +191,7 @@ def load_pretrained( progress=_DOWNLOAD_PROGRESS, check_hash=_CHECK_HASH, ) - model.load_pretrained("/home/xinchi.tian/pytorch-image-models/timm/models/efficientnetv2_t_agc-3620981a.pth") + model.load_pretrained("pytorch-image-models/timm/models/efficientnetv2_t_agc-3620981a.pth") return else: state_dict = load_state_dict_from_url( diff --git a/models/cv/classification/efficientnet_v2/ixrt/requirements.txt b/models/cv/classification/efficientnet_v2/ixrt/requirements.txt index 09e8dab9fbba81751d8862b1de0cffdbd1c1059b..e0bd6846ef7caf99a87b3be169f40bda303c8cf4 100644 --- a/models/cv/classification/efficientnet_v2/ixrt/requirements.txt +++ b/models/cv/classification/efficientnet_v2/ixrt/requirements.txt @@ -2,7 +2,7 @@ tqdm onnx onnxsim tabulate -timm==1.0.10 +timm==1.0.11 ppq pycuda protobuf==3.20.0 \ No newline at end of file diff --git a/models/cv/object_detection/hrnet/ixrt/fcos_hrnetv2p-w18-gn-head_4xb4-1x_coco.py b/models/cv/object_detection/hrnet/ixrt/fcos_hrnetv2p-w18-gn-head_4xb4-1x_coco.py index faccb7732d7063f680cf023fce4fc3ece716583d..42507ee9c23f6ed6003ef9381a180d9cc450b1dc 100644 --- a/models/cv/object_detection/hrnet/ixrt/fcos_hrnetv2p-w18-gn-head_4xb4-1x_coco.py +++ b/models/cv/object_detection/hrnet/ixrt/fcos_hrnetv2p-w18-gn-head_4xb4-1x_coco.py @@ -156,7 +156,7 @@ test_dataloader = dict( ann_file='annotations/instances_val2017.json', backend_args=None, data_prefix=dict(img='val2017/'), - data_root='/home/xinchi.tian/ixrt-modelzoo/data/datasets/coco2017', + data_root=data_root, pipeline=[ dict(backend_args=None, type='LoadImageFromFile'), dict(keep_ratio=True, scale=( @@ -181,8 +181,7 @@ test_dataloader = dict( persistent_workers=True, sampler=dict(shuffle=False, type='DefaultSampler')) test_evaluator = dict( - ann_file= - '/home/xinchi.tian/ixrt-modelzoo/data/datasets/coco2017/annotations/instances_val2017.json', + ann_file=data_root + 'annotations/image_info_test-dev2017.json', backend_args=None, format_only=False, metric='bbox',