From eeec51677a44844df0fa1d32dac3cd77187ba612 Mon Sep 17 00:00:00 2001 From: zhihang Date: Wed, 12 Mar 2025 06:57:11 +0000 Subject: [PATCH 1/7] remove all Dockerfile for OPEA --- agent-langchain/0.3.0/24.03-lts/Dockerfile | 34 ---------- agent-langchain/meta.yml | 3 - .../1.0/24.03-lts/Dockerfile | 30 --------- chatqna-conversation-ui/1.0/24.03-lts/env.sh | 15 ----- .../1.0/24.03-lts/nginx.conf | 20 ------ chatqna-conversation-ui/meta.yml | 3 - chatqna-ui/1.0/24.03-lts/Dockerfile | 31 --------- chatqna-ui/meta.yml | 2 - chatqna/1.0/24.03-lts/Dockerfile | 38 ----------- chatqna/meta.yml | 2 - codegen-ui/1.0/24.03-lts/Dockerfile | 31 --------- codegen-ui/meta.yml | 3 - codegen/1.0/24.03-lts/Dockerfile | 40 ------------ codegen/meta.yml | 3 - codetrans-ui/1.0/24.03-lts/Dockerfile | 28 -------- codetrans-ui/meta.yml | 3 - codetrans/1.0/24.03-lts/Dockerfile | 38 ----------- codetrans/meta.yml | 3 - .../0.3.4-redis5.1.1/24.03-lts/Dockerfile | 65 ------------------- dataprep-redis-langchain/meta.yml | 3 - docsum-ui/1.0/24.03-lts/Dockerfile | 31 --------- docsum-ui/meta.yml | 2 - docsum/1.0/24.03-lts/Dockerfile | 38 ----------- docsum/meta.yml | 2 - .../0.3.3/24.03-lts/Dockerfile | 48 -------------- embedding-tei-langchain/meta.yml | 2 - faqgen-ui/1.0/24.03-lts/Dockerfile | 28 -------- faqgen-ui/meta.yml | 3 - faqgen/1.0/24.03-lts/Dockerfile | 39 ----------- faqgen/meta.yml | 3 - .../0.3.0/24.03-lts/Dockerfile | 37 ----------- llm-docsum-tgi-langchain/meta.yml | 3 - .../0.3.0/24.03-lts/Dockerfile | 33 ---------- llm-faqgen-tgi-langchain/meta.yml | 3 - llm-tgi/1.0/24.03-lts/Dockerfile | 33 ---------- llm-tgi/meta.yml | 3 - .../0.3.3-npu/24.03-lts/Dockerfile | 64 ------------------ llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile | 40 ------------ llm-vllm-langchain/meta.yml | 5 -- reranking-tei/1.0/24.03-lts/Dockerfile | 49 -------------- reranking-tei/meta.yml | 2 - .../0.3.3-redis5.1.1/24.03-lts/Dockerfile | 37 ----------- retriever-redis-langchain/meta.yml | 3 - translation-ui/1.0/24.03-lts/Dockerfile | 32 --------- translation-ui/meta.yml | 3 - translation/1.0/24.03-lts/Dockerfile | 47 -------------- translation/meta.yml | 3 - 47 files changed, 988 deletions(-) delete mode 100644 agent-langchain/0.3.0/24.03-lts/Dockerfile delete mode 100644 agent-langchain/meta.yml delete mode 100644 chatqna-conversation-ui/1.0/24.03-lts/Dockerfile delete mode 100644 chatqna-conversation-ui/1.0/24.03-lts/env.sh delete mode 100644 chatqna-conversation-ui/1.0/24.03-lts/nginx.conf delete mode 100644 chatqna-conversation-ui/meta.yml delete mode 100644 chatqna-ui/1.0/24.03-lts/Dockerfile delete mode 100644 chatqna-ui/meta.yml delete mode 100644 chatqna/1.0/24.03-lts/Dockerfile delete mode 100644 chatqna/meta.yml delete mode 100644 codegen-ui/1.0/24.03-lts/Dockerfile delete mode 100644 codegen-ui/meta.yml delete mode 100644 codegen/1.0/24.03-lts/Dockerfile delete mode 100644 codegen/meta.yml delete mode 100644 codetrans-ui/1.0/24.03-lts/Dockerfile delete mode 100644 codetrans-ui/meta.yml delete mode 100644 codetrans/1.0/24.03-lts/Dockerfile delete mode 100644 codetrans/meta.yml delete mode 100644 dataprep-redis-langchain/0.3.4-redis5.1.1/24.03-lts/Dockerfile delete mode 100644 dataprep-redis-langchain/meta.yml delete mode 100644 docsum-ui/1.0/24.03-lts/Dockerfile delete mode 100644 docsum-ui/meta.yml delete mode 100644 docsum/1.0/24.03-lts/Dockerfile delete mode 100644 docsum/meta.yml delete mode 100644 embedding-tei-langchain/0.3.3/24.03-lts/Dockerfile delete mode 100644 embedding-tei-langchain/meta.yml delete mode 100644 faqgen-ui/1.0/24.03-lts/Dockerfile delete mode 100644 faqgen-ui/meta.yml delete mode 100644 faqgen/1.0/24.03-lts/Dockerfile delete mode 100644 faqgen/meta.yml delete mode 100644 llm-docsum-tgi-langchain/0.3.0/24.03-lts/Dockerfile delete mode 100644 llm-docsum-tgi-langchain/meta.yml delete mode 100644 llm-faqgen-tgi-langchain/0.3.0/24.03-lts/Dockerfile delete mode 100644 llm-faqgen-tgi-langchain/meta.yml delete mode 100644 llm-tgi/1.0/24.03-lts/Dockerfile delete mode 100644 llm-tgi/meta.yml delete mode 100644 llm-vllm-langchain/0.3.3-npu/24.03-lts/Dockerfile delete mode 100644 llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile delete mode 100644 llm-vllm-langchain/meta.yml delete mode 100644 reranking-tei/1.0/24.03-lts/Dockerfile delete mode 100644 reranking-tei/meta.yml delete mode 100644 retriever-redis-langchain/0.3.3-redis5.1.1/24.03-lts/Dockerfile delete mode 100644 retriever-redis-langchain/meta.yml delete mode 100644 translation-ui/1.0/24.03-lts/Dockerfile delete mode 100644 translation-ui/meta.yml delete mode 100644 translation/1.0/24.03-lts/Dockerfile delete mode 100644 translation/meta.yml diff --git a/agent-langchain/0.3.0/24.03-lts/Dockerfile b/agent-langchain/0.3.0/24.03-lts/Dockerfile deleted file mode 100644 index 5ccd0e1c..00000000 --- a/agent-langchain/0.3.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && \ - yum install -y \ - --setopt=install_weak_deps=False \ - python-pip python \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user -RUN git clone https://github.com/opea-project/GenAIComps.git - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/DocSum/docsum.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps - -ENTRYPOINT ["python", "docsum.py"] diff --git a/agent-langchain/meta.yml b/agent-langchain/meta.yml deleted file mode 100644 index bef603ca..00000000 --- a/agent-langchain/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: agent-langchain/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile b/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index eacfe6db..00000000 --- a/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts as vite-app - -# Update package manager and install Git -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - git \ - npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -WORKDIR /home/user/GenAIExamples/ChatQnA/ui/react - -RUN ["npm", "install"] -RUN ["npm", "run", "build"] - -FROM openeuler/nginx:1.27.2-oe2403lts - -COPY --from=vite-app /home/user/GenAIExamples/ChatQnA/ui/react/dist /usr/local/nginx/html - -COPY env.sh /docker-entrypoint.d/env.sh - -COPY nginx.conf /etc/nginx/conf.d/default.conf -RUN chmod +x /docker-entrypoint.d/env.sh \ No newline at end of file diff --git a/chatqna-conversation-ui/1.0/24.03-lts/env.sh b/chatqna-conversation-ui/1.0/24.03-lts/env.sh deleted file mode 100644 index b9660aab..00000000 --- a/chatqna-conversation-ui/1.0/24.03-lts/env.sh +++ /dev/null @@ -1,15 +0,0 @@ -#!/bin/sh -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -for i in $(env | grep APP_) #// Make sure to use the prefix MY_APP_ if you have any other prefix in env.production file variable name replace it with MY_APP_ -do - key=$(echo $i | cut -d '=' -f 1) - value=$(echo $i | cut -d '=' -f 2-) - echo $key=$value - # sed All files - # find /usr/share/nginx/html -type f -exec sed -i "s|${key}|${value}|g" '{}' + - - # sed JS and CSS only - find /usr/local/nginx/html -type f \( -name '*.js' -o -name '*.css' \) -exec sed -i "s|${key}|${value}|g" '{}' + -done diff --git a/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf b/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf deleted file mode 100644 index fa220eb7..00000000 --- a/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf +++ /dev/null @@ -1,20 +0,0 @@ -server { - listen 80; - - gzip on; - gzip_proxied any; - gzip_comp_level 6; - gzip_buffers 16 8k; - gzip_http_version 1.1; - gzip_types font/woff2 text/css application/javascript application/json application/font-woff application/font-tff image/gif image/png image/svg+xml application/octet-stream; - - location / { - root /usr/local/nginx/html; - index index.html index.htm; - try_files $uri $uri/ /index.html =404; - - location ~* \.(gif|jpe?g|png|webp|ico|svg|css|js|mp4|woff2)$ { - expires 1d; - } - } -} \ No newline at end of file diff --git a/chatqna-conversation-ui/meta.yml b/chatqna-conversation-ui/meta.yml deleted file mode 100644 index 09e4737b..00000000 --- a/chatqna-conversation-ui/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: chatqna-conversation-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/chatqna-ui/1.0/24.03-lts/Dockerfile b/chatqna-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index fb3626ad..00000000 --- a/chatqna-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - git \ - npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/ChatQnA/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/chatqna-ui/meta.yml b/chatqna-ui/meta.yml deleted file mode 100644 index 05addfa6..00000000 --- a/chatqna-ui/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -1.0-oe2403lts: - path: chatqna-ui/1.0/24.03-lts/Dockerfile \ No newline at end of file diff --git a/chatqna/1.0/24.03-lts/Dockerfile b/chatqna/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 8f3a7b81..00000000 --- a/chatqna/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,38 +0,0 @@ - - -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/ChatQnA/chatqna.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps - -USER user - -RUN echo 'ulimit -S -n 999999' >> ~/.bashrc - -ENTRYPOINT ["python", "chatqna.py"] diff --git a/chatqna/meta.yml b/chatqna/meta.yml deleted file mode 100644 index 0e837daf..00000000 --- a/chatqna/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -1.0-oe2403lts: - path: chatqna/1.0/24.03-lts/Dockerfile diff --git a/codegen-ui/1.0/24.03-lts/Dockerfile b/codegen-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 217ab697..00000000 --- a/codegen-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - git \ - npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/CodeGen/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/codegen-ui/meta.yml b/codegen-ui/meta.yml deleted file mode 100644 index 8352c985..00000000 --- a/codegen-ui/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: codegen-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/codegen/1.0/24.03-lts/Dockerfile b/codegen/1.0/24.03-lts/Dockerfile deleted file mode 100644 index fc5aac8b..00000000 --- a/codegen/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,40 +0,0 @@ - - -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ENV LANG=C.UTF-8 - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/CodeGen/codegen.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps - -USER user - -WORKDIR /home/user - -ENTRYPOINT ["python", "codegen.py"] diff --git a/codegen/meta.yml b/codegen/meta.yml deleted file mode 100644 index d9e0a7d4..00000000 --- a/codegen/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: codegen/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/codetrans-ui/1.0/24.03-lts/Dockerfile b/codetrans-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index ca3cade6..00000000 --- a/codetrans-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && yum install -y git npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/CodeTrans/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/codetrans-ui/meta.yml b/codetrans-ui/meta.yml deleted file mode 100644 index b744b447..00000000 --- a/codetrans-ui/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: codetrans-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/codetrans/1.0/24.03-lts/Dockerfile b/codetrans/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 3f45677c..00000000 --- a/codetrans/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,38 +0,0 @@ - - -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/CodeTrans/code_translation.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps - -USER user - -WORKDIR /home/user - -ENTRYPOINT ["python", "code_translation.py"] diff --git a/codetrans/meta.yml b/codetrans/meta.yml deleted file mode 100644 index 6985ae55..00000000 --- a/codetrans/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: codetrans/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/dataprep-redis-langchain/0.3.4-redis5.1.1/24.03-lts/Dockerfile b/dataprep-redis-langchain/0.3.4-redis5.1.1/24.03-lts/Dockerfile deleted file mode 100644 index 73d7c51c..00000000 --- a/dataprep-redis-langchain/0.3.4-redis5.1.1/24.03-lts/Dockerfile +++ /dev/null @@ -1,65 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ENV LANG=C.UTF-8 - -ARG ARCH="cpu" - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - wget \ - java-11-openjdk \ - mesa-libGL \ - shadow-utils \ - jemalloc jemalloc-devel \ - poppler-utils \ - python-pip python \ - tesseract \ - tesseract-langpack-eng \ - git - - -RUN mkdir -p /tmp/LibreOffice/LibreOffice && \ - wget -O /tmp/LibreOffice/LibreOffice.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/24.8.5/rpm/x86_64/LibreOffice_24.8.5_Linux_x86-64_rpm.tar.gz && \ - tar -zxvf /tmp/LibreOffice/LibreOffice.tar.gz -C /tmp/LibreOffice/LibreOffice/ --strip-components 1 && \ - yum -y install /tmp/LibreOffice/LibreOffice/RPMS/*.rpm && \ - mkdir -p /tmp/LibreOffice/langpack_zh-CN && \ - wget -O /tmp/LibreOffice/langpack_zh-CN.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/24.8.5/rpm/x86_64/LibreOffice_24.8.5_Linux_x86-64_rpm_langpack_zh-CN.tar.gz && \ - tar -zxvf /tmp/LibreOffice/langpack_zh-CN.tar.gz -C /tmp/LibreOffice/langpack_zh-CN/ --strip-components 1 && \ - yum -y install /tmp/LibreOffice/langpack_zh-CN/RPMS/*.rpm && \ - mkdir -p /tmp/LibreOffice/helppack_zh-CN && \ - wget -O /tmp/LibreOffice/helppack_zh-CN.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/24.8.5/rpm/x86_64/LibreOffice_24.8.5_Linux_x86-64_rpm_helppack_zh-CN.tar.gz && \ - tar -zxvf /tmp/LibreOffice/helppack_zh-CN.tar.gz -C /tmp/LibreOffice/helppack_zh-CN/ --strip-components 1 && \ - yum -y install /tmp/LibreOffice/helppack_zh-CN/RPMS/*.rpm && \ - rm -rf /tmp/LibreOffice - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip setuptools && \ - if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ - pip install --no-cache-dir -r /home/user/comps/dataprep/redis/langchain/requirements.txt - -ENV PYTHONPATH=/usr/bin/python:/home/user - -USER root - -RUN mkdir -p /home/user/comps/dataprep/redis/langchain/uploaded_files && chown -R user /home/user/comps/dataprep/redis/langchain/uploaded_files - -USER user - -WORKDIR /home/user/comps/dataprep/redis/langchain - -ENTRYPOINT ["python", "prepare_doc_redis.py"] \ No newline at end of file diff --git a/dataprep-redis-langchain/meta.yml b/dataprep-redis-langchain/meta.yml deleted file mode 100644 index 079bce74..00000000 --- a/dataprep-redis-langchain/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -0.3.4-redis5.1.1-oe2403lts: - path: dataprep-redis-langchain/0.3.4-redis5.1.1/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/docsum-ui/1.0/24.03-lts/Dockerfile b/docsum-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 278871d5..00000000 --- a/docsum-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - git \ - npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/DocSum/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/docsum-ui/meta.yml b/docsum-ui/meta.yml deleted file mode 100644 index bd1d176b..00000000 --- a/docsum-ui/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -1.0-oe2403lts: - path: docsum-ui/1.0/24.03-lts/Dockerfile diff --git a/docsum/1.0/24.03-lts/Dockerfile b/docsum/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 8be51f87..00000000 --- a/docsum/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,38 +0,0 @@ - - -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -WORKDIR /home/user -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/DocSum/docsum.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps - -USER user - -WORKDIR /home/user - -ENTRYPOINT ["python", "docsum.py"] diff --git a/docsum/meta.yml b/docsum/meta.yml deleted file mode 100644 index 51e1f507..00000000 --- a/docsum/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -1.0-oe2403lts: - path: docsum/1.0/24.03-lts/Dockerfile diff --git a/embedding-tei-langchain/0.3.3/24.03-lts/Dockerfile b/embedding-tei-langchain/0.3.3/24.03-lts/Dockerfile deleted file mode 100644 index 9e3e6fc5..00000000 --- a/embedding-tei-langchain/0.3.3/24.03-lts/Dockerfile +++ /dev/null @@ -1,48 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 -FROM openeuler/openeuler:24.03-lts - -ARG ARCH="cpu" - -RUN yum update -y && \ - yum install -y \ - --setopt=install_weak_deps=False \ - python3-pip python3-devel \ - cmake gcc-c++ git \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ - CFLAGS="-I /tmp/vhacdx" pip3 install . && \ - rm -rf /tmp/vhacdx - -RUN pip3 install scikit-build-core numpy && \ - cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ - cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ - pip3 install --upgrade pip setuptools wheel build nanobind ninja && \ - pip install "nanobind>=1.8.0,<=2.2.0" && \ - python3 -m build --no-isolation --sdist --wheel && \ - pip3 install dist/manifold3d-*.whl - -WORKDIR /home/user/ - -RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ - cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -RUN if [ ${ARCH} = "cpu" ]; then pip3 install --no-cache-dir torch --index-url https://download.pytorch.org/whl/cpu; fi && \ -pip3 install --no-cache-dir -r /home/user/comps/embeddings/tei/langchain/requirements.txt - -ENV PYTHONPATH=/usr/bin/python3:/home/user - -WORKDIR /home/user/comps/embeddings/tei/langchain - -ENTRYPOINT ["python3", "embedding_tei.py"] diff --git a/embedding-tei-langchain/meta.yml b/embedding-tei-langchain/meta.yml deleted file mode 100644 index 1192bf09..00000000 --- a/embedding-tei-langchain/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -0.3.3-oe2403lts: - path: embedding-tei-langchain/0.3.3/24.03-lts/Dockerfile \ No newline at end of file diff --git a/faqgen-ui/1.0/24.03-lts/Dockerfile b/faqgen-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 84139439..00000000 --- a/faqgen-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && yum install -y git npm - -WORKDIR /home/user - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/FaqGen/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/faqgen-ui/meta.yml b/faqgen-ui/meta.yml deleted file mode 100644 index a9ddecf5..00000000 --- a/faqgen-ui/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: faqgen-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/faqgen/1.0/24.03-lts/Dockerfile b/faqgen/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 411d8bfb..00000000 --- a/faqgen/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,39 +0,0 @@ - - -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - - -RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -RUN cp GenAIExamples/FaqGen/faqgen.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps - -USER user - -WORKDIR /home/user - -ENTRYPOINT ["python", "faqgen.py"] diff --git a/faqgen/meta.yml b/faqgen/meta.yml deleted file mode 100644 index 385daeb3..00000000 --- a/faqgen/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: faqgen/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/llm-docsum-tgi-langchain/0.3.0/24.03-lts/Dockerfile b/llm-docsum-tgi-langchain/0.3.0/24.03-lts/Dockerfile deleted file mode 100644 index 5665f56b..00000000 --- a/llm-docsum-tgi-langchain/0.3.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ARG PLATFORM="cpu" - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip && \ - if [ ${PLATFORM} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ - pip install --no-cache-dir -r /home/user/comps/llms/summarization/tgi/langchain/requirements.txt - -ENV PYTHONPATH=$PYTHONPATH:/home/user - -WORKDIR /home/user/comps/llms/summarization/tgi/langchain - -ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/llm-docsum-tgi-langchain/meta.yml b/llm-docsum-tgi-langchain/meta.yml deleted file mode 100644 index c0e12e35..00000000 --- a/llm-docsum-tgi-langchain/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -0.3.0-oe2403lts: - path: llm-docsum-tgi-langchain/0.3.0/24.03-lts/Dockerfile - arch: x86_64 diff --git a/llm-faqgen-tgi-langchain/0.3.0/24.03-lts/Dockerfile b/llm-faqgen-tgi-langchain/0.3.0/24.03-lts/Dockerfile deleted file mode 100644 index 279a1ef9..00000000 --- a/llm-faqgen-tgi-langchain/0.3.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,33 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip setuptools && \ - pip install --no-cache-dir -r /home/user/comps/llms/faq-generation/tgi/langchain/requirements.txt - -ENV PYTHONPATH=$PYTHONPATH:/home/user - -WORKDIR /home/user/comps/llms/faq-generation/tgi/langchain - -ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/llm-faqgen-tgi-langchain/meta.yml b/llm-faqgen-tgi-langchain/meta.yml deleted file mode 100644 index 2ac0967c..00000000 --- a/llm-faqgen-tgi-langchain/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -0.3.0-oe2403lts: - path: llm-faqgen-tgi-langchain/0.3.0/24.03-lts/Dockerfile - arch: x86_64 diff --git a/llm-tgi/1.0/24.03-lts/Dockerfile b/llm-tgi/1.0/24.03-lts/Dockerfile deleted file mode 100644 index b39bcc81..00000000 --- a/llm-tgi/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,33 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip setuptools && \ - pip install --no-cache-dir -r /home/user/comps/llms/text-generation/tgi/requirements.txt - -ENV PYTHONPATH=/usr/bin/python:/home/user - -WORKDIR /home/user/comps/llms/text-generation/tgi - -ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/llm-tgi/meta.yml b/llm-tgi/meta.yml deleted file mode 100644 index 5d5b47f7..00000000 --- a/llm-tgi/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: llm-tgi/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/llm-vllm-langchain/0.3.3-npu/24.03-lts/Dockerfile b/llm-vllm-langchain/0.3.3-npu/24.03-lts/Dockerfile deleted file mode 100644 index 5b01ae74..00000000 --- a/llm-vllm-langchain/0.3.3-npu/24.03-lts/Dockerfile +++ /dev/null @@ -1,64 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 -FROM openeuler/openeuler:22.03-lts - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python python3-devel \ - cmake gcc-c++ \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ - cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ - CFLAGS="-I /tmp/vhacdx" pip install . && \ - rm -rf /tmp/vhacdx - -RUN python -m pip install scikit-build-core numpy && \ - cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ - cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ - python -m pip install --upgrade pip setuptools wheel build nanobind ninja && \ - python -m build --no-isolation --sdist --wheel && \ - python -m pip install dist/manifold3d-*.whl && \ - rm -rf /tmp/manifold - -RUN python3 -m pip install --no-cache-dir --upgrade pip && \ - python3 -m pip install --no-cache-dir \ - fastapi \ - huggingface_hub \ - langchain \ - langchain-huggingface \ - langchain-openai \ - langchain_community \ - langchainhub \ - opentelemetry-api \ - opentelemetry-exporter-otlp \ - opentelemetry-sdk \ - prometheus-fastapi-instrumentator \ - shortuuid \ - transformers \ - uvicorn - -RUN pip3 install docarray[full] -RUN git clone -b npu_support --depth 1 https://github.com/wangshuai09/vllm.git -RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" python3 -m pip install -r ./vllm/requirements-build.txt -RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" VLLM_TARGET_DEVICE="npu" python3 -m pip install ./vllm/ && \ - rm -rf ./vllm - -ENV PYTHONPATH=/usr/bin/python:/home/user -WORKDIR /home/user/comps/llms/text-generation/vllm/langchain - -ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile b/llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile deleted file mode 100644 index 4dd93f48..00000000 --- a/llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile +++ /dev/null @@ -1,40 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ARG ARCH="cpu" # Set this to "cpu" or "gpu" - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip && \ - if [ ${ARCH} = "cpu" ]; then \ - pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \ - else \ - pip install --no-cache-dir -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \ - fi - -ENV PYTHONPATH=/usr/bin/python:/home/user - -WORKDIR /home/user/comps/llms/text-generation/vllm/langchain - -ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/llm-vllm-langchain/meta.yml b/llm-vllm-langchain/meta.yml deleted file mode 100644 index 003354fc..00000000 --- a/llm-vllm-langchain/meta.yml +++ /dev/null @@ -1,5 +0,0 @@ -0.3.3-oe2403lts: - path: llm-vllm-langchain/0.3.3/24.03-lts/Dockerfile - arch: x86_64 -0.3.3-npu-oe2203lts: - path: llm-vllm-langchain/0.3.3-npu/24.03-lts/Dockerfile \ No newline at end of file diff --git a/reranking-tei/1.0/24.03-lts/Dockerfile b/reranking-tei/1.0/24.03-lts/Dockerfile deleted file mode 100644 index ee09dd44..00000000 --- a/reranking-tei/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,49 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ARG ARCH="cpu" - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python python3-devel \ - cmake gcc-c++ git \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ - CFLAGS="-I /tmp/vhacdx" pip install . && \ - rm -rf /tmp/vhacdx - -RUN python -m pip install scikit-build-core numpy && \ - cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ - cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ - python -m pip install --upgrade pip setuptools wheel build nanobind ninja && \ - python -m build --no-isolation --sdist --wheel && \ - python -m pip install dist/manifold3d-*.whl && \ - rm -rf /tmp/manifold - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ - cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip && \ - if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch --index-url https://download.pytorch.org/whl/cpu; fi && \ - pip install --no-cache-dir -r /home/user/comps/reranks/tei/requirements.txt - -ENV PYTHONPATH=/usr/bin/python:/home/user - -WORKDIR /home/user/comps/reranks/tei - -ENTRYPOINT ["python", "reranking_tei.py"] diff --git a/reranking-tei/meta.yml b/reranking-tei/meta.yml deleted file mode 100644 index 3423455e..00000000 --- a/reranking-tei/meta.yml +++ /dev/null @@ -1,2 +0,0 @@ -1.0-oe2403lts: - path: reranking-tei/1.0/24.03-lts/Dockerfile \ No newline at end of file diff --git a/retriever-redis-langchain/0.3.3-redis5.1.1/24.03-lts/Dockerfile b/retriever-redis-langchain/0.3.3-redis5.1.1/24.03-lts/Dockerfile deleted file mode 100644 index ca061bb6..00000000 --- a/retriever-redis-langchain/0.3.3-redis5.1.1/24.03-lts/Dockerfile +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -FROM openeuler/openeuler:24.03-lts - -ARG ARCH="cpu" - -RUN yum update -y && \ - yum install -y --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ - -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -RUN cp -r GenAIComps/comps . && \ - rm -rf GenAIComps - -RUN pip install --no-cache-dir --upgrade pip && \ - if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ - pip install --no-cache-dir -r /home/user/comps/retrievers/redis/langchain/requirements.txt - -ENV PYTHONPATH=/usr/bin/python:/home/user - -WORKDIR /home/user/comps/retrievers/redis/langchain - -ENTRYPOINT ["python", "retriever_redis.py"] diff --git a/retriever-redis-langchain/meta.yml b/retriever-redis-langchain/meta.yml deleted file mode 100644 index f07b1cb6..00000000 --- a/retriever-redis-langchain/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -0.3.3-redis5.1.1-oe2403lts: - path: retriever-redis-langchain/0.3.3-redis5.1.1/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/translation-ui/1.0/24.03-lts/Dockerfile b/translation-ui/1.0/24.03-lts/Dockerfile deleted file mode 100644 index c41e5086..00000000 --- a/translation-ui/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -# Use node 20.11.1 as the base image -FROM openeuler/openeuler:24.03-lts - -# Update package manager and install Git -RUN yum update -y && \ - yum install -y \ - --setopt=install_weak_deps=False \ - npm \ - git - -WORKDIR /home/user/ - -# Copy the front-end code repository -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 - -# Set the working directory -WORKDIR /home/user/GenAIExamples/Translation/ui/svelte - -# Install front-end dependencies -RUN npm install - -# Build the front-end application -RUN npm run build - -# Expose the port of the front-end application -EXPOSE 5173 - -# Run the front-end application in preview mode -CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] diff --git a/translation-ui/meta.yml b/translation-ui/meta.yml deleted file mode 100644 index fded9031..00000000 --- a/translation-ui/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: translation-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file diff --git a/translation/1.0/24.03-lts/Dockerfile b/translation/1.0/24.03-lts/Dockerfile deleted file mode 100644 index 493eb7bb..00000000 --- a/translation/1.0/24.03-lts/Dockerfile +++ /dev/null @@ -1,47 +0,0 @@ -# Copyright (c) 2024 Intel Corporation -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - - -FROM openeuler/openeuler:24.03-lts - -RUN yum update -y && \ - yum install -y \ - --setopt=install_weak_deps=False \ - python-pip python \ - shadow-utils \ - mesa-libGL \ - jemalloc-devel \ - git - -RUN useradd -m -s /bin/bash user && \ - mkdir -p /home/user && \ - chown -R user /home/user/ - -USER user - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 - -WORKDIR /home/user/GenAIComps -RUN pip install --no-cache-dir --upgrade pip && \ - pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt - -WORKDIR /home/user/ -RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 -RUN cp GenAIExamples/Translation/translation.py . -RUN rm -rf GenAIExamples - -ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps - -ENTRYPOINT ["python", "translation.py"] diff --git a/translation/meta.yml b/translation/meta.yml deleted file mode 100644 index ec4a3517..00000000 --- a/translation/meta.yml +++ /dev/null @@ -1,3 +0,0 @@ -1.0-oe2403lts: - path: translation/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file -- Gitee From ba53f02c24398d75954f836f9644825bb649bc37 Mon Sep 17 00:00:00 2001 From: zhihang Date: Thu, 13 Mar 2025 02:21:53 +0000 Subject: [PATCH 2/7] update OPEA --- .../1.0/24.03-lts/Dockerfile | 30 +++++++++ .../1.0/24.03-lts/env.sh | 15 +++++ .../1.0/24.03-lts/nginx.conf | 20 ++++++ opea/chatqna-conversation-ui/meta.yml | 3 + opea/chatqna-ui/1.0/24.03-lts/Dockerfile | 31 +++++++++ opea/chatqna-ui/meta.yml | 3 + opea/chatqna/1.0/24.03-lts/Dockerfile | 38 +++++++++++ opea/chatqna/meta.yml | 3 + opea/dataprep-redis/1.0/24.03-lts/Dockerfile | 65 +++++++++++++++++++ opea/dataprep-redis/meta.yml | 3 + opea/embedding-tei/1.0/24.03-lts/Dockerfile | 48 ++++++++++++++ opea/embedding-tei/meta.yml | 3 + opea/llm-tgi/1.0/24.03-lts/Dockerfile | 33 ++++++++++ opea/llm-tgi/meta.yml | 3 + opea/reranking-tei/1.0/24.03-lts/Dockerfile | 49 ++++++++++++++ opea/reranking-tei/meta.yml | 3 + opea/retriever-redis/1.0/24.03-lts/Dockerfile | 37 +++++++++++ opea/retriever-redis/meta.yml | 3 + 18 files changed, 390 insertions(+) create mode 100644 opea/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/chatqna-conversation-ui/1.0/24.03-lts/env.sh create mode 100644 opea/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf create mode 100644 opea/chatqna-conversation-ui/meta.yml create mode 100644 opea/chatqna-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/chatqna-ui/meta.yml create mode 100644 opea/chatqna/1.0/24.03-lts/Dockerfile create mode 100644 opea/chatqna/meta.yml create mode 100644 opea/dataprep-redis/1.0/24.03-lts/Dockerfile create mode 100644 opea/dataprep-redis/meta.yml create mode 100644 opea/embedding-tei/1.0/24.03-lts/Dockerfile create mode 100644 opea/embedding-tei/meta.yml create mode 100644 opea/llm-tgi/1.0/24.03-lts/Dockerfile create mode 100644 opea/llm-tgi/meta.yml create mode 100644 opea/reranking-tei/1.0/24.03-lts/Dockerfile create mode 100644 opea/reranking-tei/meta.yml create mode 100644 opea/retriever-redis/1.0/24.03-lts/Dockerfile create mode 100644 opea/retriever-redis/meta.yml diff --git a/opea/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile b/opea/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..eacfe6db --- /dev/null +++ b/opea/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,30 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts as vite-app + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + git \ + npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +WORKDIR /home/user/GenAIExamples/ChatQnA/ui/react + +RUN ["npm", "install"] +RUN ["npm", "run", "build"] + +FROM openeuler/nginx:1.27.2-oe2403lts + +COPY --from=vite-app /home/user/GenAIExamples/ChatQnA/ui/react/dist /usr/local/nginx/html + +COPY env.sh /docker-entrypoint.d/env.sh + +COPY nginx.conf /etc/nginx/conf.d/default.conf +RUN chmod +x /docker-entrypoint.d/env.sh \ No newline at end of file diff --git a/opea/chatqna-conversation-ui/1.0/24.03-lts/env.sh b/opea/chatqna-conversation-ui/1.0/24.03-lts/env.sh new file mode 100644 index 00000000..b9660aab --- /dev/null +++ b/opea/chatqna-conversation-ui/1.0/24.03-lts/env.sh @@ -0,0 +1,15 @@ +#!/bin/sh +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +for i in $(env | grep APP_) #// Make sure to use the prefix MY_APP_ if you have any other prefix in env.production file variable name replace it with MY_APP_ +do + key=$(echo $i | cut -d '=' -f 1) + value=$(echo $i | cut -d '=' -f 2-) + echo $key=$value + # sed All files + # find /usr/share/nginx/html -type f -exec sed -i "s|${key}|${value}|g" '{}' + + + # sed JS and CSS only + find /usr/local/nginx/html -type f \( -name '*.js' -o -name '*.css' \) -exec sed -i "s|${key}|${value}|g" '{}' + +done diff --git a/opea/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf b/opea/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf new file mode 100644 index 00000000..fa220eb7 --- /dev/null +++ b/opea/chatqna-conversation-ui/1.0/24.03-lts/nginx.conf @@ -0,0 +1,20 @@ +server { + listen 80; + + gzip on; + gzip_proxied any; + gzip_comp_level 6; + gzip_buffers 16 8k; + gzip_http_version 1.1; + gzip_types font/woff2 text/css application/javascript application/json application/font-woff application/font-tff image/gif image/png image/svg+xml application/octet-stream; + + location / { + root /usr/local/nginx/html; + index index.html index.htm; + try_files $uri $uri/ /index.html =404; + + location ~* \.(gif|jpe?g|png|webp|ico|svg|css|js|mp4|woff2)$ { + expires 1d; + } + } +} \ No newline at end of file diff --git a/opea/chatqna-conversation-ui/meta.yml b/opea/chatqna-conversation-ui/meta.yml new file mode 100644 index 00000000..09e4737b --- /dev/null +++ b/opea/chatqna-conversation-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: chatqna-conversation-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/chatqna-ui/1.0/24.03-lts/Dockerfile b/opea/chatqna-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..fb3626ad --- /dev/null +++ b/opea/chatqna-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,31 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + git \ + npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/ChatQnA/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/chatqna-ui/meta.yml b/opea/chatqna-ui/meta.yml new file mode 100644 index 00000000..e3709ef9 --- /dev/null +++ b/opea/chatqna-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: chatqna-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/chatqna/1.0/24.03-lts/Dockerfile b/opea/chatqna/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..8f3a7b81 --- /dev/null +++ b/opea/chatqna/1.0/24.03-lts/Dockerfile @@ -0,0 +1,38 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/ChatQnA/chatqna.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps + +USER user + +RUN echo 'ulimit -S -n 999999' >> ~/.bashrc + +ENTRYPOINT ["python", "chatqna.py"] diff --git a/opea/chatqna/meta.yml b/opea/chatqna/meta.yml new file mode 100644 index 00000000..db51af04 --- /dev/null +++ b/opea/chatqna/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: chatqna/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/dataprep-redis/1.0/24.03-lts/Dockerfile b/opea/dataprep-redis/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..3ca47f9f --- /dev/null +++ b/opea/dataprep-redis/1.0/24.03-lts/Dockerfile @@ -0,0 +1,65 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ENV LANG=C.UTF-8 + +ARG ARCH="cpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + wget \ + java-11-openjdk \ + mesa-libGL \ + shadow-utils \ + jemalloc jemalloc-devel \ + poppler-utils \ + python-pip python \ + tesseract \ + tesseract-langpack-eng \ + git + + +RUN mkdir -p /tmp/LibreOffice/LibreOffice && \ + wget -O /tmp/LibreOffice/LibreOffice.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/25.2.1/rpm/x86_64/LibreOffice_25.2.1_Linux_x86-64_rpm.tar.gz && \ + tar -zxvf /tmp/LibreOffice/LibreOffice.tar.gz -C /tmp/LibreOffice/LibreOffice/ --strip-components 1 && \ + yum -y install /tmp/LibreOffice/LibreOffice/RPMS/*.rpm && \ + mkdir -p /tmp/LibreOffice/langpack_zh-CN && \ + wget -O /tmp/LibreOffice/langpack_zh-CN.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/25.2.1/rpm/x86_64/LibreOffice_25.2.1_Linux_x86-64_rpm_langpack_zh-CN.tar.gz && \ + tar -zxvf /tmp/LibreOffice/langpack_zh-CN.tar.gz -C /tmp/LibreOffice/langpack_zh-CN/ --strip-components 1 && \ + yum -y install /tmp/LibreOffice/langpack_zh-CN/RPMS/*.rpm && \ + mkdir -p /tmp/LibreOffice/helppack_zh-CN && \ + wget -O /tmp/LibreOffice/helppack_zh-CN.tar.gz https://mirrors.cloud.tencent.com/libreoffice/libreoffice/stable/25.2.1/rpm/x86_64/LibreOffice_25.2.1_Linux_x86-64_rpm_helppack_zh-CN.tar.gz && \ + tar -zxvf /tmp/LibreOffice/helppack_zh-CN.tar.gz -C /tmp/LibreOffice/helppack_zh-CN/ --strip-components 1 && \ + yum -y install /tmp/LibreOffice/helppack_zh-CN/RPMS/*.rpm && \ + rm -rf /tmp/LibreOffice + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip setuptools && \ + if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ + pip install --no-cache-dir -r /home/user/comps/dataprep/redis/langchain/requirements.txt + +ENV PYTHONPATH=/usr/bin/python:/home/user + +USER root + +RUN mkdir -p /home/user/comps/dataprep/redis/langchain/uploaded_files && chown -R user /home/user/comps/dataprep/redis/langchain/uploaded_files + +USER user + +WORKDIR /home/user/comps/dataprep/redis/langchain + +ENTRYPOINT ["python", "prepare_doc_redis.py"] diff --git a/opea/dataprep-redis/meta.yml b/opea/dataprep-redis/meta.yml new file mode 100644 index 00000000..58cdff22 --- /dev/null +++ b/opea/dataprep-redis/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: dataprep-redis/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/embedding-tei/1.0/24.03-lts/Dockerfile b/opea/embedding-tei/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..aa862333 --- /dev/null +++ b/opea/embedding-tei/1.0/24.03-lts/Dockerfile @@ -0,0 +1,48 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" + +RUN yum update -y && \ + yum install -y \ + --setopt=install_weak_deps=False \ + python3-pip python3-devel \ + cmake gcc-c++ \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ + CFLAGS="-I /tmp/vhacdx" pip3 install . && \ + rm -rf /tmp/vhacdx + +RUN pip3 install scikit-build-core numpy && \ + cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ + cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ + pip3 install --upgrade pip setuptools wheel build nanobind ninja && \ + pip install "nanobind>=1.8.0,<=2.2.0" && \ + python3 -m build --no-isolation --sdist --wheel && \ + pip3 install dist/manifold3d-*.whl + +WORKDIR /home/user/ + +RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ + cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +RUN if [ ${ARCH} = "cpu" ]; then pip3 install --no-cache-dir torch --index-url https://download.pytorch.org/whl/cpu; fi && \ +pip3 install --no-cache-dir -r /home/user/comps/embeddings/tei/langchain/requirements.txt + +ENV PYTHONPATH=/usr/bin/python3:/home/user + +WORKDIR /home/user/comps/embeddings/tei/langchain + +ENTRYPOINT ["python3", "embedding_tei.py"] diff --git a/opea/embedding-tei/meta.yml b/opea/embedding-tei/meta.yml new file mode 100644 index 00000000..f2f8cf17 --- /dev/null +++ b/opea/embedding-tei/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: embedding-tei/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/llm-tgi/1.0/24.03-lts/Dockerfile b/opea/llm-tgi/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..b39bcc81 --- /dev/null +++ b/opea/llm-tgi/1.0/24.03-lts/Dockerfile @@ -0,0 +1,33 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip setuptools && \ + pip install --no-cache-dir -r /home/user/comps/llms/text-generation/tgi/requirements.txt + +ENV PYTHONPATH=/usr/bin/python:/home/user + +WORKDIR /home/user/comps/llms/text-generation/tgi + +ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/opea/llm-tgi/meta.yml b/opea/llm-tgi/meta.yml new file mode 100644 index 00000000..5d5b47f7 --- /dev/null +++ b/opea/llm-tgi/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: llm-tgi/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/reranking-tei/1.0/24.03-lts/Dockerfile b/opea/reranking-tei/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..ee09dd44 --- /dev/null +++ b/opea/reranking-tei/1.0/24.03-lts/Dockerfile @@ -0,0 +1,49 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python python3-devel \ + cmake gcc-c++ git \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ + CFLAGS="-I /tmp/vhacdx" pip install . && \ + rm -rf /tmp/vhacdx + +RUN python -m pip install scikit-build-core numpy && \ + cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ + cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ + python -m pip install --upgrade pip setuptools wheel build nanobind ninja && \ + python -m build --no-isolation --sdist --wheel && \ + python -m pip install dist/manifold3d-*.whl && \ + rm -rf /tmp/manifold + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ + cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch --index-url https://download.pytorch.org/whl/cpu; fi && \ + pip install --no-cache-dir -r /home/user/comps/reranks/tei/requirements.txt + +ENV PYTHONPATH=/usr/bin/python:/home/user + +WORKDIR /home/user/comps/reranks/tei + +ENTRYPOINT ["python", "reranking_tei.py"] diff --git a/opea/reranking-tei/meta.yml b/opea/reranking-tei/meta.yml new file mode 100644 index 00000000..39a54523 --- /dev/null +++ b/opea/reranking-tei/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: reranking-tei/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/retriever-redis/1.0/24.03-lts/Dockerfile b/opea/retriever-redis/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..ca061bb6 --- /dev/null +++ b/opea/retriever-redis/1.0/24.03-lts/Dockerfile @@ -0,0 +1,37 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ + pip install --no-cache-dir -r /home/user/comps/retrievers/redis/langchain/requirements.txt + +ENV PYTHONPATH=/usr/bin/python:/home/user + +WORKDIR /home/user/comps/retrievers/redis/langchain + +ENTRYPOINT ["python", "retriever_redis.py"] diff --git a/opea/retriever-redis/meta.yml b/opea/retriever-redis/meta.yml new file mode 100644 index 00000000..69d2c62b --- /dev/null +++ b/opea/retriever-redis/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: retriever-redis/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file -- Gitee From 8f5142266981d4208b19eb9e1d2f24b31792834b Mon Sep 17 00:00:00 2001 From: zhihang Date: Thu, 13 Mar 2025 02:56:20 +0000 Subject: [PATCH 3/7] update meta.yml --- opea/chatqna-conversation-ui/meta.yml | 4 ++-- opea/chatqna-ui/meta.yml | 4 ++-- opea/chatqna/meta.yml | 2 +- opea/dataprep-redis/meta.yml | 4 ++-- opea/embedding-tei/meta.yml | 4 ++-- opea/llm-tgi/meta.yml | 4 ++-- opea/reranking-tei/meta.yml | 4 ++-- opea/retriever-redis/meta.yml | 4 ++-- 8 files changed, 15 insertions(+), 15 deletions(-) diff --git a/opea/chatqna-conversation-ui/meta.yml b/opea/chatqna-conversation-ui/meta.yml index 09e4737b..bd5103c7 100644 --- a/opea/chatqna-conversation-ui/meta.yml +++ b/opea/chatqna-conversation-ui/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: chatqna-conversation-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/chatqna-conversation-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/chatqna-ui/meta.yml b/opea/chatqna-ui/meta.yml index e3709ef9..9cec7c11 100644 --- a/opea/chatqna-ui/meta.yml +++ b/opea/chatqna-ui/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: chatqna-ui/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/chatqna-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/chatqna/meta.yml b/opea/chatqna/meta.yml index db51af04..34843ee1 100644 --- a/opea/chatqna/meta.yml +++ b/opea/chatqna/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: chatqna/1.0/24.03-lts/Dockerfile + path: opea/chatqna/1.0/24.03-lts/Dockerfile arch: x86_64 diff --git a/opea/dataprep-redis/meta.yml b/opea/dataprep-redis/meta.yml index 58cdff22..7dade52e 100644 --- a/opea/dataprep-redis/meta.yml +++ b/opea/dataprep-redis/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: dataprep-redis/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/dataprep-redis/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/embedding-tei/meta.yml b/opea/embedding-tei/meta.yml index f2f8cf17..be89b811 100644 --- a/opea/embedding-tei/meta.yml +++ b/opea/embedding-tei/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: embedding-tei/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/embedding-tei/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/llm-tgi/meta.yml b/opea/llm-tgi/meta.yml index 5d5b47f7..e718271f 100644 --- a/opea/llm-tgi/meta.yml +++ b/opea/llm-tgi/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: llm-tgi/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/llm-tgi/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/reranking-tei/meta.yml b/opea/reranking-tei/meta.yml index 39a54523..2af46ef4 100644 --- a/opea/reranking-tei/meta.yml +++ b/opea/reranking-tei/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: reranking-tei/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/reranking-tei/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/retriever-redis/meta.yml b/opea/retriever-redis/meta.yml index 69d2c62b..6ceb61aa 100644 --- a/opea/retriever-redis/meta.yml +++ b/opea/retriever-redis/meta.yml @@ -1,3 +1,3 @@ 1.0-oe2403lts: - path: retriever-redis/1.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + path: opea/retriever-redis/1.0/24.03-lts/Dockerfile + arch: x86_64 -- Gitee From b4ec6d781e95bdecd71a7d498d5f35a480b367d9 Mon Sep 17 00:00:00 2001 From: zhihang Date: Fri, 14 Mar 2025 14:51:48 +0000 Subject: [PATCH 4/7] Update OPEA images --- opea/codegen-ui/1.0/24.03-lts/Dockerfile | 31 ++++++++++++ opea/codegen-ui/meta.yml | 3 ++ opea/codegen/1.0/24.03-lts/Dockerfile | 40 ++++++++++++++++ opea/codegen/meta.yml | 3 ++ opea/codetrans-ui/1.0/24.03-lts/Dockerfile | 28 +++++++++++ opea/codetrans-ui/meta.yml | 3 ++ opea/codetrans/1.0/24.03-lts/Dockerfile | 38 +++++++++++++++ opea/codetrans/meta.yml | 3 ++ opea/docsum-ui/1.0/24.03-lts/Dockerfile | 31 ++++++++++++ opea/docsum-ui/meta.yml | 3 ++ opea/docsum/1.0/24.03-lts/Dockerfile | 38 +++++++++++++++ opea/docsum/meta.yml | 3 ++ opea/faqgen-ui/1.0/24.03-lts/Dockerfile | 28 +++++++++++ opea/faqgen-ui/meta.yml | 3 ++ opea/faqgen/1.0/24.03-lts/Dockerfile | 39 +++++++++++++++ opea/faqgen/meta.yml | 3 ++ opea/llm-docsum-tgi/1.0/24.03-lts/Dockerfile | 37 +++++++++++++++ opea/llm-docsum-tgi/meta.yml | 3 ++ opea/llm-faqgen-tgi/1.0/24.03-lts/Dockerfile | 33 +++++++++++++ opea/llm-faqgen-tgi/meta.yml | 3 ++ opea/translation-ui/1.0/24.03-lts/Dockerfile | 32 +++++++++++++ opea/translation-ui/meta.yml | 3 ++ opea/translation/1.0/24.03-lts/Dockerfile | 47 +++++++++++++++++++ opea/translation/meta.yml | 3 ++ .../{1.5 => 1.5.0}/24.03-lts/Dockerfile | 0 text-embeddings-inference-cpu/meta.yml | 6 +-- text-generation-inference-cpu/meta.yml | 2 +- 27 files changed, 462 insertions(+), 4 deletions(-) create mode 100644 opea/codegen-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/codegen-ui/meta.yml create mode 100644 opea/codegen/1.0/24.03-lts/Dockerfile create mode 100644 opea/codegen/meta.yml create mode 100644 opea/codetrans-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/codetrans-ui/meta.yml create mode 100644 opea/codetrans/1.0/24.03-lts/Dockerfile create mode 100644 opea/codetrans/meta.yml create mode 100644 opea/docsum-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/docsum-ui/meta.yml create mode 100644 opea/docsum/1.0/24.03-lts/Dockerfile create mode 100644 opea/docsum/meta.yml create mode 100644 opea/faqgen-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/faqgen-ui/meta.yml create mode 100644 opea/faqgen/1.0/24.03-lts/Dockerfile create mode 100644 opea/faqgen/meta.yml create mode 100644 opea/llm-docsum-tgi/1.0/24.03-lts/Dockerfile create mode 100644 opea/llm-docsum-tgi/meta.yml create mode 100644 opea/llm-faqgen-tgi/1.0/24.03-lts/Dockerfile create mode 100644 opea/llm-faqgen-tgi/meta.yml create mode 100644 opea/translation-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/translation-ui/meta.yml create mode 100644 opea/translation/1.0/24.03-lts/Dockerfile create mode 100644 opea/translation/meta.yml rename text-embeddings-inference-cpu/{1.5 => 1.5.0}/24.03-lts/Dockerfile (100%) diff --git a/opea/codegen-ui/1.0/24.03-lts/Dockerfile b/opea/codegen-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..217ab697 --- /dev/null +++ b/opea/codegen-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,31 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + git \ + npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/CodeGen/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/codegen-ui/meta.yml b/opea/codegen-ui/meta.yml new file mode 100644 index 00000000..8352c985 --- /dev/null +++ b/opea/codegen-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: codegen-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/codegen/1.0/24.03-lts/Dockerfile b/opea/codegen/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..fc5aac8b --- /dev/null +++ b/opea/codegen/1.0/24.03-lts/Dockerfile @@ -0,0 +1,40 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ENV LANG=C.UTF-8 + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/CodeGen/codegen.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "codegen.py"] diff --git a/opea/codegen/meta.yml b/opea/codegen/meta.yml new file mode 100644 index 00000000..d9e0a7d4 --- /dev/null +++ b/opea/codegen/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: codegen/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/codetrans-ui/1.0/24.03-lts/Dockerfile b/opea/codetrans-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..ca3cade6 --- /dev/null +++ b/opea/codetrans-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,28 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && yum install -y git npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/CodeTrans/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/codetrans-ui/meta.yml b/opea/codetrans-ui/meta.yml new file mode 100644 index 00000000..b744b447 --- /dev/null +++ b/opea/codetrans-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: codetrans-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/codetrans/1.0/24.03-lts/Dockerfile b/opea/codetrans/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..3f45677c --- /dev/null +++ b/opea/codetrans/1.0/24.03-lts/Dockerfile @@ -0,0 +1,38 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/CodeTrans/code_translation.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=/usr/bin/python:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "code_translation.py"] diff --git a/opea/codetrans/meta.yml b/opea/codetrans/meta.yml new file mode 100644 index 00000000..6985ae55 --- /dev/null +++ b/opea/codetrans/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: codetrans/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/docsum-ui/1.0/24.03-lts/Dockerfile b/opea/docsum-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..278871d5 --- /dev/null +++ b/opea/docsum-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,31 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + git \ + npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/DocSum/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/docsum-ui/meta.yml b/opea/docsum-ui/meta.yml new file mode 100644 index 00000000..40975cbf --- /dev/null +++ b/opea/docsum-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: docsum-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/docsum/1.0/24.03-lts/Dockerfile b/opea/docsum/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..8be51f87 --- /dev/null +++ b/opea/docsum/1.0/24.03-lts/Dockerfile @@ -0,0 +1,38 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/DocSum/docsum.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "docsum.py"] diff --git a/opea/docsum/meta.yml b/opea/docsum/meta.yml new file mode 100644 index 00000000..4d78dccc --- /dev/null +++ b/opea/docsum/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: docsum/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/faqgen-ui/1.0/24.03-lts/Dockerfile b/opea/faqgen-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..84139439 --- /dev/null +++ b/opea/faqgen-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,28 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && yum install -y git npm + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/FaqGen/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/faqgen-ui/meta.yml b/opea/faqgen-ui/meta.yml new file mode 100644 index 00000000..a9ddecf5 --- /dev/null +++ b/opea/faqgen-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: faqgen-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/faqgen/1.0/24.03-lts/Dockerfile b/opea/faqgen/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..411d8bfb --- /dev/null +++ b/opea/faqgen/1.0/24.03-lts/Dockerfile @@ -0,0 +1,39 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + + +RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +RUN cp GenAIExamples/FaqGen/faqgen.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "faqgen.py"] diff --git a/opea/faqgen/meta.yml b/opea/faqgen/meta.yml new file mode 100644 index 00000000..385daeb3 --- /dev/null +++ b/opea/faqgen/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: faqgen/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/llm-docsum-tgi/1.0/24.03-lts/Dockerfile b/opea/llm-docsum-tgi/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..10a54537 --- /dev/null +++ b/opea/llm-docsum-tgi/1.0/24.03-lts/Dockerfile @@ -0,0 +1,37 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \ + pip install --no-cache-dir -r /home/user/comps/llms/summarization/tgi/langchain/requirements.txt + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/llms/summarization/tgi/langchain + +ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/opea/llm-docsum-tgi/meta.yml b/opea/llm-docsum-tgi/meta.yml new file mode 100644 index 00000000..9fdd5888 --- /dev/null +++ b/opea/llm-docsum-tgi/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: llm-docsum-tgi/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/llm-faqgen-tgi/1.0/24.03-lts/Dockerfile b/opea/llm-faqgen-tgi/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..279a1ef9 --- /dev/null +++ b/opea/llm-faqgen-tgi/1.0/24.03-lts/Dockerfile @@ -0,0 +1,33 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip setuptools && \ + pip install --no-cache-dir -r /home/user/comps/llms/faq-generation/tgi/langchain/requirements.txt + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/llms/faq-generation/tgi/langchain + +ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/opea/llm-faqgen-tgi/meta.yml b/opea/llm-faqgen-tgi/meta.yml new file mode 100644 index 00000000..7bf7d9b2 --- /dev/null +++ b/opea/llm-faqgen-tgi/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: llm-faqgen-tgi/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/translation-ui/1.0/24.03-lts/Dockerfile b/opea/translation-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..c41e5086 --- /dev/null +++ b/opea/translation-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,32 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y \ + --setopt=install_weak_deps=False \ + npm \ + git + +WORKDIR /home/user/ + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/Translation/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] diff --git a/opea/translation-ui/meta.yml b/opea/translation-ui/meta.yml new file mode 100644 index 00000000..fded9031 --- /dev/null +++ b/opea/translation-ui/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: translation-ui/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/translation/1.0/24.03-lts/Dockerfile b/opea/translation/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..493eb7bb --- /dev/null +++ b/opea/translation/1.0/24.03-lts/Dockerfile @@ -0,0 +1,47 @@ +# Copyright (c) 2024 Intel Corporation +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y \ + --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/Translation/translation.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +ENTRYPOINT ["python", "translation.py"] diff --git a/opea/translation/meta.yml b/opea/translation/meta.yml new file mode 100644 index 00000000..ec4a3517 --- /dev/null +++ b/opea/translation/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: translation/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/text-embeddings-inference-cpu/1.5/24.03-lts/Dockerfile b/text-embeddings-inference-cpu/1.5.0/24.03-lts/Dockerfile similarity index 100% rename from text-embeddings-inference-cpu/1.5/24.03-lts/Dockerfile rename to text-embeddings-inference-cpu/1.5.0/24.03-lts/Dockerfile diff --git a/text-embeddings-inference-cpu/meta.yml b/text-embeddings-inference-cpu/meta.yml index fa1a239e..7eb2a7ae 100644 --- a/text-embeddings-inference-cpu/meta.yml +++ b/text-embeddings-inference-cpu/meta.yml @@ -1,3 +1,3 @@ -1.5-oe2403lts: - path: text-embeddings-inference-cpu/1.5/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file +1.5.0-oe2403lts: + path: text-embeddings-inference-cpu/1.5.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/text-generation-inference-cpu/meta.yml b/text-generation-inference-cpu/meta.yml index f2d6d770..f525aa77 100644 --- a/text-generation-inference-cpu/meta.yml +++ b/text-generation-inference-cpu/meta.yml @@ -1,3 +1,3 @@ 2.4.0-oe2403lts: path: text-generation-inference-cpu/2.4.0/24.03-lts/Dockerfile - arch: x86_64 \ No newline at end of file + arch: x86_64 -- Gitee From bf94f8ea609885620754d27640907f45563f480e Mon Sep 17 00:00:00 2001 From: zhihang Date: Tue, 18 Mar 2025 07:24:47 +0000 Subject: [PATCH 5/7] Reapply "add openEuler support for SearchQnA" This reverts commit 5612c1b0e87a4d1ac874e99d124e9bbe1a7f4f44. --- opea/asr/1.0/24.03-LTS/Dockerfile | 39 +++++++++++ opea/asr/meta.yml | 3 + opea/audioqna/1.0/24.03-LTS/Dockerfile | 38 +++++++++++ opea/audioqna/meta.yml | 3 + .../1.0/24.03-LTS/Dockerfile | 34 ++++++++++ opea/doc-index-retriever/meta.yml | 3 + opea/finetuning/1.0/24.03-LTS/Dockerfile | 49 ++++++++++++++ opea/finetuning/meta.yml | 3 + opea/llm-vllm/0.3.3-npu/24.03-lts/Dockerfile | 64 +++++++++++++++++++ opea/llm-vllm/1.0/24.03-lts/Dockerfile | 40 ++++++++++++ opea/llm-vllm/meta.yml | 3 + opea/searchqna-ui/1.0/24.03-lts/Dockerfile | 31 +++++++++ opea/searchqna-ui/meta.yml | 2 + opea/searchqna/1.0/24.03-lts/Dockerfile | 36 +++++++++++ opea/searchqna/meta.yml | 2 + opea/speecht5/1.0/24.03-LTS/Dockerfile | 44 +++++++++++++ opea/speecht5/meta.yml | 3 + opea/tts/1.0/24.03-LTS/Dockerfile | 39 +++++++++++ opea/tts/meta.yml | 3 + .../1.0/24.03-lts/Dockerfile | 40 ++++++++++++ opea/web-retriever-chroma/meta.yml | 3 + opea/whisper/1.0/24.03-LTS/Dockerfile | 44 +++++++++++++ opea/whisper/meta.yml | 3 + python/3.9.9/24.03-lts/Dockerfile | 62 ++++++++++++++++++ 24 files changed, 591 insertions(+) create mode 100644 opea/asr/1.0/24.03-LTS/Dockerfile create mode 100644 opea/asr/meta.yml create mode 100644 opea/audioqna/1.0/24.03-LTS/Dockerfile create mode 100644 opea/audioqna/meta.yml create mode 100644 opea/doc-index-retriever/1.0/24.03-LTS/Dockerfile create mode 100644 opea/doc-index-retriever/meta.yml create mode 100644 opea/finetuning/1.0/24.03-LTS/Dockerfile create mode 100644 opea/finetuning/meta.yml create mode 100644 opea/llm-vllm/0.3.3-npu/24.03-lts/Dockerfile create mode 100644 opea/llm-vllm/1.0/24.03-lts/Dockerfile create mode 100644 opea/llm-vllm/meta.yml create mode 100644 opea/searchqna-ui/1.0/24.03-lts/Dockerfile create mode 100644 opea/searchqna-ui/meta.yml create mode 100644 opea/searchqna/1.0/24.03-lts/Dockerfile create mode 100644 opea/searchqna/meta.yml create mode 100644 opea/speecht5/1.0/24.03-LTS/Dockerfile create mode 100644 opea/speecht5/meta.yml create mode 100644 opea/tts/1.0/24.03-LTS/Dockerfile create mode 100644 opea/tts/meta.yml create mode 100644 opea/web-retriever-chroma/1.0/24.03-lts/Dockerfile create mode 100644 opea/web-retriever-chroma/meta.yml create mode 100644 opea/whisper/1.0/24.03-LTS/Dockerfile create mode 100644 opea/whisper/meta.yml create mode 100644 python/3.9.9/24.03-lts/Dockerfile diff --git a/opea/asr/1.0/24.03-LTS/Dockerfile b/opea/asr/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..b3a4c361 --- /dev/null +++ b/opea/asr/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,39 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ +USER user + +ENV LANG=C.UTF-8 +ARG ARCH=cpu + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ "${ARCH}" = "cpu" ]; then \ + pip install torch torchvision --index-url https://download.pytorch.org/whl/cpu ; \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/asr/whisper/requirements.txt ; \ + else \ + pip install --no-cache-dir -r /home/user/comps/asr/whisper/requirements.txt ; \ + fi + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/asr/whisper + +ENTRYPOINT ["python", "asr.py"] \ No newline at end of file diff --git a/opea/asr/meta.yml b/opea/asr/meta.yml new file mode 100644 index 00000000..86aea5c3 --- /dev/null +++ b/opea/asr/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: asr/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/audioqna/1.0/24.03-LTS/Dockerfile b/opea/audioqna/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..d84616c0 --- /dev/null +++ b/opea/audioqna/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,38 @@ + + +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/AudioQnA/audioqna.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "audioqna.py"] diff --git a/opea/audioqna/meta.yml b/opea/audioqna/meta.yml new file mode 100644 index 00000000..6aaab5ac --- /dev/null +++ b/opea/audioqna/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: audioqna/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/doc-index-retriever/1.0/24.03-LTS/Dockerfile b/opea/doc-index-retriever/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..e127024b --- /dev/null +++ b/opea/doc-index-retriever/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,34 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/DocIndexRetriever/retrieval_tool.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +ENTRYPOINT ["python", "retrieval_tool.py"] diff --git a/opea/doc-index-retriever/meta.yml b/opea/doc-index-retriever/meta.yml new file mode 100644 index 00000000..7b3bbed0 --- /dev/null +++ b/opea/doc-index-retriever/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: doc-index-retriever/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/finetuning/1.0/24.03-LTS/Dockerfile b/opea/finetuning/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..d1ca9ae4 --- /dev/null +++ b/opea/finetuning/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,49 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use the same python version with ray +FROM openeuler/openeuler:24.03-lts + +ARG HF_TOKEN + +ENV HF_TOKEN=$HF_TOKEN + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +ENV PATH=$PATH:/home/user/.local/bin + +RUN python -m pip install --no-cache-dir --upgrade pip && \ + python -m pip install --no-cache-dir torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu && \ + python -m pip install --no-cache-dir intel-extension-for-pytorch && \ + python -m pip install --no-cache-dir oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ + python -m pip install --no-cache-dir -r /home/user/comps/finetuning/requirements.txt + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/finetuning + +RUN echo PKGPATH=$(python3 -c "import pkg_resources; print(pkg_resources.get_distribution('oneccl-bind-pt').location)") >> run.sh && \ + echo 'export LD_LIBRARY_PATH=$PKGPATH/oneccl_bindings_for_pytorch/opt/mpi/lib/:$LD_LIBRARY_PATH' >> run.sh && \ + echo 'source $PKGPATH/oneccl_bindings_for_pytorch/env/setvars.sh' >> run.sh && \ + echo ray start --head --dashboard-host=0.0.0.0 >> run.sh && \ + echo export RAY_ADDRESS=http://localhost:8265 >> run.sh && \ + echo python finetuning_service.py >> run.sh + +CMD bash run.sh diff --git a/opea/finetuning/meta.yml b/opea/finetuning/meta.yml new file mode 100644 index 00000000..e71455ef --- /dev/null +++ b/opea/finetuning/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: finetuning/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/llm-vllm/0.3.3-npu/24.03-lts/Dockerfile b/opea/llm-vllm/0.3.3-npu/24.03-lts/Dockerfile new file mode 100644 index 00000000..5b01ae74 --- /dev/null +++ b/opea/llm-vllm/0.3.3-npu/24.03-lts/Dockerfile @@ -0,0 +1,64 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 +FROM openeuler/openeuler:22.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python python3-devel \ + cmake gcc-c++ \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone --depth=1 --branch v1.0 https://github.com/opea-project/GenAIComps.git && \ + cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN cd /tmp && git clone --depth 1 https://github.com/trimesh/vhacdx && cd vhacdx && \ + CFLAGS="-I /tmp/vhacdx" pip install . && \ + rm -rf /tmp/vhacdx + +RUN python -m pip install scikit-build-core numpy && \ + cd /tmp && git clone --depth 1 https://github.com/elalish/manifold.git && \ + cd manifold && git clone --depth 1 --branch v2021.10.0 https://github.com/oneapi-src/oneTBB.git && \ + python -m pip install --upgrade pip setuptools wheel build nanobind ninja && \ + python -m build --no-isolation --sdist --wheel && \ + python -m pip install dist/manifold3d-*.whl && \ + rm -rf /tmp/manifold + +RUN python3 -m pip install --no-cache-dir --upgrade pip && \ + python3 -m pip install --no-cache-dir \ + fastapi \ + huggingface_hub \ + langchain \ + langchain-huggingface \ + langchain-openai \ + langchain_community \ + langchainhub \ + opentelemetry-api \ + opentelemetry-exporter-otlp \ + opentelemetry-sdk \ + prometheus-fastapi-instrumentator \ + shortuuid \ + transformers \ + uvicorn + +RUN pip3 install docarray[full] +RUN git clone -b npu_support --depth 1 https://github.com/wangshuai09/vllm.git +RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" python3 -m pip install -r ./vllm/requirements-build.txt +RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" VLLM_TARGET_DEVICE="npu" python3 -m pip install ./vllm/ && \ + rm -rf ./vllm + +ENV PYTHONPATH=/usr/bin/python:/home/user +WORKDIR /home/user/comps/llms/text-generation/vllm/langchain + +ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/opea/llm-vllm/1.0/24.03-lts/Dockerfile b/opea/llm-vllm/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..4dd93f48 --- /dev/null +++ b/opea/llm-vllm/1.0/24.03-lts/Dockerfile @@ -0,0 +1,40 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" # Set this to "cpu" or "gpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ ${ARCH} = "cpu" ]; then \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \ + else \ + pip install --no-cache-dir -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \ + fi + +ENV PYTHONPATH=/usr/bin/python:/home/user + +WORKDIR /home/user/comps/llms/text-generation/vllm/langchain + +ENTRYPOINT ["bash", "entrypoint.sh"] diff --git a/opea/llm-vllm/meta.yml b/opea/llm-vllm/meta.yml new file mode 100644 index 00000000..9d48c8de --- /dev/null +++ b/opea/llm-vllm/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: llm-vllm/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/searchqna-ui/1.0/24.03-lts/Dockerfile b/opea/searchqna-ui/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..2b2d4222 --- /dev/null +++ b/opea/searchqna-ui/1.0/24.03-lts/Dockerfile @@ -0,0 +1,31 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +# Use node 20.11.1 as the base image +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + npm \ + git + +WORKDIR /home/user + +# Copy the front-end code repository +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 + +# Set the working directory +WORKDIR /home/user/GenAIExamples/SearchQnA/ui/svelte + +# Install front-end dependencies +RUN npm install + +# Build the front-end application +RUN npm run build + +# Expose the port of the front-end application +EXPOSE 5173 + +# Run the front-end application in preview mode +CMD ["npm", "run", "preview", "--", "--port", "5173", "--host", "0.0.0.0"] \ No newline at end of file diff --git a/opea/searchqna-ui/meta.yml b/opea/searchqna-ui/meta.yml new file mode 100644 index 00000000..5ffeedf0 --- /dev/null +++ b/opea/searchqna-ui/meta.yml @@ -0,0 +1,2 @@ +1.0-oe2403lts: + path: searchqna-ui/1.0/24.03-lts/Dockerfile diff --git a/opea/searchqna/1.0/24.03-lts/Dockerfile b/opea/searchqna/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..b77d3b5f --- /dev/null +++ b/opea/searchqna/1.0/24.03-lts/Dockerfile @@ -0,0 +1,36 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +WORKDIR /home/user/GenAIComps +RUN pip install --no-cache-dir --upgrade pip && \ + pip install --no-cache-dir -r /home/user/GenAIComps/requirements.txt + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIExamples.git && cd GenAIExamples && git checkout v1.0 +RUN cp GenAIExamples/SearchQnA/searchqna.py . +RUN rm -rf GenAIExamples + +ENV PYTHONPATH=$PYTHONPATH:/home/user/GenAIComps + +USER user + +WORKDIR /home/user + +ENTRYPOINT ["python", "searchqna.py"] diff --git a/opea/searchqna/meta.yml b/opea/searchqna/meta.yml new file mode 100644 index 00000000..fede55bd --- /dev/null +++ b/opea/searchqna/meta.yml @@ -0,0 +1,2 @@ +1.0-oe2403lts: + path: searchqna/1.0/24.03-lts/Dockerfile diff --git a/opea/speecht5/1.0/24.03-LTS/Dockerfile b/opea/speecht5/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..44066319 --- /dev/null +++ b/opea/speecht5/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,44 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +# Install system dependencies +RUN yum update -y && \ + yum install -y \ + --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + ffmpeg \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +# Set environment variables +ENV LANG=en_US.UTF-8 +ENV PYTHONPATH=/home/user +ARG ARCH=cpu + +USER user + +WORKDIR /home/user +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.2 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip setuptools && \ + if [ "${ARCH}" = "cpu" ]; then \ + pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu ; \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/tts/src/integrations/dependency/speecht5/requirements.txt ; \ + else \ + pip install --no-cache-dir -r /home/user/comps/tts/src/integrations/dependency/speecht5/requirements.txt ; \ + fi + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/tts/src/integrations/dependency/speecht5 + +ENTRYPOINT ["python", "speecht5_server.py", "--device", "cpu"] diff --git a/opea/speecht5/meta.yml b/opea/speecht5/meta.yml new file mode 100644 index 00000000..65e48466 --- /dev/null +++ b/opea/speecht5/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: speecht5/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/tts/1.0/24.03-LTS/Dockerfile b/opea/tts/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..60ccfa28 --- /dev/null +++ b/opea/tts/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,39 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +# Update package manager and install Git +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ +USER user + +ENV LANG=C.UTF-8 +ARG ARCH=cpu + +WORKDIR /home/user/ +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ "${ARCH}" = "cpu" ]; then \ + pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu ; \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/tts/speecht5/requirements.txt ; \ + else \ + pip install --no-cache-dir -r /home/user/comps/tts/speecht5/requirements.txt ; \ + fi + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/tts/speecht5 + +ENTRYPOINT ["python", "tts.py"] diff --git a/opea/tts/meta.yml b/opea/tts/meta.yml new file mode 100644 index 00000000..57d18eeb --- /dev/null +++ b/opea/tts/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: tts/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/opea/web-retriever-chroma/1.0/24.03-lts/Dockerfile b/opea/web-retriever-chroma/1.0/24.03-lts/Dockerfile new file mode 100644 index 00000000..61056432 --- /dev/null +++ b/opea/web-retriever-chroma/1.0/24.03-lts/Dockerfile @@ -0,0 +1,40 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +ARG ARCH="cpu" # Set this to "cpu" or "gpu" + +RUN yum update -y && \ + yum install -y --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + mesa-libGL \ + jemalloc-devel \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +USER user + +WORKDIR /home/user/ + +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.0 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip && \ + if [ ${ARCH} = "cpu" ]; then \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/web_retrievers/chroma/langchain/requirements.txt; \ + else \ + pip install --no-cache-dir -r /home/user/comps/web_retrievers/chroma/langchain/requirements.txt; \ + fi + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/web_retrievers/chroma/langchain + +ENTRYPOINT ["python", "retriever_chroma.py"] diff --git a/opea/web-retriever-chroma/meta.yml b/opea/web-retriever-chroma/meta.yml new file mode 100644 index 00000000..6b18568d --- /dev/null +++ b/opea/web-retriever-chroma/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: web-retriever-chroma/1.0/24.03-lts/Dockerfile + arch: x86_64 diff --git a/opea/whisper/1.0/24.03-LTS/Dockerfile b/opea/whisper/1.0/24.03-LTS/Dockerfile new file mode 100644 index 00000000..2a29d8b8 --- /dev/null +++ b/opea/whisper/1.0/24.03-LTS/Dockerfile @@ -0,0 +1,44 @@ +# Copyright (C) 2024 Intel Corporation +# SPDX-License-Identifier: Apache-2.0 + +FROM openeuler/openeuler:24.03-lts + +# Install system dependencies +RUN yum update -y && \ + yum install -y \ + --setopt=install_weak_deps=False \ + python-pip python \ + shadow-utils \ + ffmpeg \ + git + +RUN useradd -m -s /bin/bash user && \ + mkdir -p /home/user && \ + chown -R user /home/user/ + +# Set environment variables +ENV LANG=en_US.UTF-8 +ARG ARCH=cpu + +USER user + +WORKDIR /home/user +RUN git clone https://github.com/opea-project/GenAIComps.git && cd GenAIComps && git checkout v1.2 + +RUN cp -r GenAIComps/comps . && \ + rm -rf GenAIComps + +RUN pip install --no-cache-dir --upgrade pip setuptools && \ + pip install --no-cache-dir -r /home/user/comps/asr/src/requirements.txt && \ + if [ "${ARCH}" = "cpu" ]; then \ + pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu ; \ + pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/asr/src/requirements.txt ; \ + else \ + pip install --no-cache-dir -r /home/user/comps/asr/src/requirements.txt ; \ + fi + +ENV PYTHONPATH=$PYTHONPATH:/home/user + +WORKDIR /home/user/comps/asr/src/integrations/dependency/whisper + +ENTRYPOINT ["python", "whisper_server.py", "--device", "cpu"] diff --git a/opea/whisper/meta.yml b/opea/whisper/meta.yml new file mode 100644 index 00000000..3f43ec8b --- /dev/null +++ b/opea/whisper/meta.yml @@ -0,0 +1,3 @@ +1.0-oe2403lts: + path: whisper/1.0/24.03-lts/Dockerfile + arch: x86_64 \ No newline at end of file diff --git a/python/3.9.9/24.03-lts/Dockerfile b/python/3.9.9/24.03-lts/Dockerfile new file mode 100644 index 00000000..1aa72049 --- /dev/null +++ b/python/3.9.9/24.03-lts/Dockerfile @@ -0,0 +1,62 @@ +FROM openeuler/openeuler:24.03-lts AS py-installer + +ARG PY_VERSION=3.9 +ARG PY_MAJOR_VERSION=3 +ARG PY_LATEST_VERSION=3.9.9 +ARG PY_HOME="/usr/local/python${PY_VERSION}" +ARG PY_INSTALLER_TGZ="Python-${PY_LATEST_VERSION}.tgz" +ARG PY_INSTALLER_DIR="Python-${PY_LATEST_VERSION}" +ARG PY_INSTALLER_URL="https://repo.huaweicloud.com/python/${PY_LATEST_VERSION}/${PY_INSTALLER_TGZ}" + +# Install dependencies +RUN yum update -y && \ + yum install -y \ + gcc \ + gcc-c++ \ + make \ + cmake \ + curl \ + zlib-devel \ + bzip2-devel \ + openssl-devel \ + ncurses-devel \ + sqlite-devel \ + readline-devel \ + tk-devel \ + gdbm-devel \ + libpcap-devel \ + xz-devel \ + libev-devel \ + expat-devel \ + libffi-devel \ + systemtap-sdt-devel \ + && yum clean all \ + && rm -rf /var/cache/yum \ + && rm -rf /tmp/* + +RUN curl -fsSL -o "/tmp/${PY_INSTALLER_TGZ}" "${PY_INSTALLER_URL}" && \ + tar -xf /tmp/${PY_INSTALLER_TGZ} -C /tmp && \ + cd /tmp/${PY_INSTALLER_DIR} && \ + mkdir -p /${PY_HOME}/lib && \ + ./configure --prefix=${PY_HOME} --enable-shared LDFLAGS="-Wl,-rpath ${PY_HOME}/lib" && \ + make -j "$(nproc)" && \ + make altinstall + + +RUN rm -rf /tmp/${PY_INSTALLER_TGZ} /tmp/${PY_INSTALLER_DIR} + +FROM openeuler/openeuler:24.03-lts AS official + +ARG PY_VERSION=3.9 +ARG PY_MAJOR_VERSION=3 +ARG PY_HOME="/usr/local/python${PY_VERSION}" + +COPY --from=py-installer /usr/local/python${PY_VERSION} /usr/local/python${PY_VERSION} + +RUN ln -sf ${PY_HOME}/bin/python${PY_VERSION} /usr/bin/python${PY_VERSION} && \ + ln -sf ${PY_HOME}/bin/pip${PY_VERSION} /usr/bin/pip${PY_VERSION} && \ + ln -sf /usr/bin/python${PY_VERSION} /usr/bin/python${PY_MAJOR_VERSION} && \ + ln -sf /usr/bin/pip${PY_VERSION} /usr/bin/pip${PY_MAJOR_VERSION} && \ + ln -sf /usr/bin/python${PY_MAJOR_VERSION} /usr/bin/python && \ + ln -sf /usr/bin/pip${PY_MAJOR_VERSION} /usr/bin/pip + -- Gitee From 3bc240e1ca0a60fdf9edc836f4972d79e7ae633e Mon Sep 17 00:00:00 2001 From: zhihang Date: Tue, 18 Mar 2025 07:40:24 +0000 Subject: [PATCH 6/7] add image-list.yml --- opea/image-list.yml | 30 ++++++++++++++++++++++++++++++ 1 file changed, 30 insertions(+) create mode 100644 opea/image-list.yml diff --git a/opea/image-list.yml b/opea/image-list.yml new file mode 100644 index 00000000..8a852bf0 --- /dev/null +++ b/opea/image-list.yml @@ -0,0 +1,30 @@ +images: + asr: opea/asr + audioqna: opea/audioqna + chatqna-conversation-ui: opea/chatqna-conversation-ui + chatqna-ui: opea/chatqna-ui + chatqna: opea/chatqna + codegen-ui: opea/codegen-ui + codegen: opea/codegen + codetrans-ui: opea/codetrans-ui + codetrans: opea/codetrans + dataprep-redis: opea/dataprep-redis + doc-index-retriever: opea/doc-index-retriever + docsum-ui: opea/docsum-ui + docsum: opea/docsum + embedding-tei: opea/embedding-tei + faqgen-ui: opea/faqgen-ui + faqgen: opea/faqgen + llm-docsum-tgi: opea/llm-docsum-tgi + llm-faqgen-tgi: opea/llm-faqgen-tgi + llm-tgi: opea/llm-tgi + reranking-tei: opea/ai/reranking-tei + retriever-redis: opea/retriever-redis + searchqna: opea/searchqna + searchqna-ui: opea/searchqna-ui + speecht5: opea/speecht5 + translation-ui: opea/translation-ui + translation: opea/translation + tts: opea/tts + web-retriever-chroma: opea/web-retriever-chroma + whisper: opea/whisper -- Gitee From a235f45d3954196c997597438004aad1cc924e7c Mon Sep 17 00:00:00 2001 From: zhihang Date: Tue, 18 Mar 2025 08:40:27 +0000 Subject: [PATCH 7/7] update image-list.yml --- opea/image-list.yml | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/opea/image-list.yml b/opea/image-list.yml index 8a852bf0..842406e1 100644 --- a/opea/image-list.yml +++ b/opea/image-list.yml @@ -18,7 +18,8 @@ images: llm-docsum-tgi: opea/llm-docsum-tgi llm-faqgen-tgi: opea/llm-faqgen-tgi llm-tgi: opea/llm-tgi - reranking-tei: opea/ai/reranking-tei + llm-vllm: opea/llm-vllm + reranking-tei: opea/reranking-tei retriever-redis: opea/retriever-redis searchqna: opea/searchqna searchqna-ui: opea/searchqna-ui @@ -28,3 +29,5 @@ images: tts: opea/tts web-retriever-chroma: opea/web-retriever-chroma whisper: opea/whisper + finetuning: opea/finetuning + -- Gitee