diff --git a/frameworks/transformers/5.0.0/Dockerfile b/frameworks/transformers/5.0.0/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..078c16a8f2e4273a776e1bb8e772e7be69a99385 --- /dev/null +++ b/frameworks/transformers/5.0.0/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.0.0 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.0.0 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.0.0/README.md b/frameworks/transformers/5.0.0/README.md new file mode 100644 index 0000000000000000000000000000000000000000..12410139180c05d298c45b4f1b082ddaed140b92 --- /dev/null +++ b/frameworks/transformers/5.0.0/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.0.0 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.0.0 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.0.0 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.0.0 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.0.0 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.0.0/build.conf b/frameworks/transformers/5.0.0/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..cabba525dded9eac72bd21041b78f685be63f6c6 --- /dev/null +++ b/frameworks/transformers/5.0.0/build.conf @@ -0,0 +1,4 @@ +# transformers 5.0.0 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.0.0 +GPU_TEST=true diff --git a/frameworks/transformers/5.0.0/test.sh b/frameworks/transformers/5.0.0/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..bcca08c38704a33265187b12c1725b491d4d917d --- /dev/null +++ b/frameworks/transformers/5.0.0/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.0.0 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.0.0', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.0.0/test_result.png b/frameworks/transformers/5.0.0/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..231060a0b9e1e5eab0e4a352accaf900f6c94f92 Binary files /dev/null and b/frameworks/transformers/5.0.0/test_result.png differ diff --git a/frameworks/transformers/5.1.0/Dockerfile b/frameworks/transformers/5.1.0/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..b525b121dbe6d9dd7742f10564eda4b457560ddf --- /dev/null +++ b/frameworks/transformers/5.1.0/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.1.0 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.1.0 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.1.0/README.md b/frameworks/transformers/5.1.0/README.md new file mode 100644 index 0000000000000000000000000000000000000000..48d1aadc4a1e03ee2c81624132a9a946f9987506 --- /dev/null +++ b/frameworks/transformers/5.1.0/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.1.0 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.1.0 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.1.0 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.1.0 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.1.0 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.1.0/build.conf b/frameworks/transformers/5.1.0/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..be843368bf308900b5f40c95b854dd22131359e6 --- /dev/null +++ b/frameworks/transformers/5.1.0/build.conf @@ -0,0 +1,4 @@ +# transformers 5.1.0 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.1.0 +GPU_TEST=true diff --git a/frameworks/transformers/5.1.0/test.sh b/frameworks/transformers/5.1.0/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..d2429cf463539b36c66319bc2061e8f48f38d055 --- /dev/null +++ b/frameworks/transformers/5.1.0/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.1.0 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.1.0', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.1.0/test_result.png b/frameworks/transformers/5.1.0/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..a2c16a5fa965eb42a378077ab53e988cb4cff82c Binary files /dev/null and b/frameworks/transformers/5.1.0/test_result.png differ diff --git a/frameworks/transformers/5.2.0/Dockerfile b/frameworks/transformers/5.2.0/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..e36675b6e32744fc3ec69c485ba5bc026bc5efa7 --- /dev/null +++ b/frameworks/transformers/5.2.0/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.2.0 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.2.0 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.2.0/README.md b/frameworks/transformers/5.2.0/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e87147ff8f684b770406d86f0959fc72aaa733c5 --- /dev/null +++ b/frameworks/transformers/5.2.0/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.2.0 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.2.0 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.2.0 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.2.0 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.2.0 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.2.0/build.conf b/frameworks/transformers/5.2.0/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..0b08cd66874a2821037925f749af2d0d770c1c00 --- /dev/null +++ b/frameworks/transformers/5.2.0/build.conf @@ -0,0 +1,4 @@ +# transformers 5.2.0 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.2.0 +GPU_TEST=true diff --git a/frameworks/transformers/5.2.0/test.sh b/frameworks/transformers/5.2.0/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..13a24543c59ef3bd5e45f272028189def5eb4c41 --- /dev/null +++ b/frameworks/transformers/5.2.0/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.2.0 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.2.0', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.2.0/test_result.png b/frameworks/transformers/5.2.0/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..cb1fe6dad6e32d188bca96e7d482e243965bf09d Binary files /dev/null and b/frameworks/transformers/5.2.0/test_result.png differ diff --git a/frameworks/transformers/5.3.0/Dockerfile b/frameworks/transformers/5.3.0/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..968934b0827ddfd3996407b729e6e93144cbf50c --- /dev/null +++ b/frameworks/transformers/5.3.0/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.3.0 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.3.0 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.3.0/README.md b/frameworks/transformers/5.3.0/README.md new file mode 100644 index 0000000000000000000000000000000000000000..5c95ff3e6a6e69f96cc61ca0c1f1fa93b1b14103 --- /dev/null +++ b/frameworks/transformers/5.3.0/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.3.0 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.3.0 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.3.0 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.3.0 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.3.0 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.3.0/build.conf b/frameworks/transformers/5.3.0/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..dcc451eec291f7909d6490e6a84a4a658fb137eb --- /dev/null +++ b/frameworks/transformers/5.3.0/build.conf @@ -0,0 +1,4 @@ +# transformers 5.3.0 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.3.0 +GPU_TEST=true diff --git a/frameworks/transformers/5.3.0/test.sh b/frameworks/transformers/5.3.0/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..55755e1a5345331a0290641ffae43fb8afa3eaf3 --- /dev/null +++ b/frameworks/transformers/5.3.0/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.3.0 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.3.0', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.3.0/test_result.png b/frameworks/transformers/5.3.0/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..8f509257264466d0a919edba835f3d7dc03d2bd6 Binary files /dev/null and b/frameworks/transformers/5.3.0/test_result.png differ diff --git a/frameworks/transformers/5.4.0/Dockerfile b/frameworks/transformers/5.4.0/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..8620d99a3a09bbe0a6f8ce8c4fc6d5bd4d7d8be4 --- /dev/null +++ b/frameworks/transformers/5.4.0/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.4.0 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.4.0 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.4.0/README.md b/frameworks/transformers/5.4.0/README.md new file mode 100644 index 0000000000000000000000000000000000000000..eb041eb365144696782be3826fbcef8fbf44f486 --- /dev/null +++ b/frameworks/transformers/5.4.0/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.4.0 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.4.0 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.4.0 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.4.0 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.4.0 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.4.0/build.conf b/frameworks/transformers/5.4.0/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..ea84be4514431666de75939bf6e7c38a0aa3c10d --- /dev/null +++ b/frameworks/transformers/5.4.0/build.conf @@ -0,0 +1,4 @@ +# transformers 5.4.0 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.4.0 +GPU_TEST=true diff --git a/frameworks/transformers/5.4.0/test.sh b/frameworks/transformers/5.4.0/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..6bf0313618953473c1e5af688425c59815ac236f --- /dev/null +++ b/frameworks/transformers/5.4.0/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.4.0 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.4.0', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.4.0/test_result.png b/frameworks/transformers/5.4.0/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..b6accfa8e7b43a18452e5f26d49cddc39496faa3 Binary files /dev/null and b/frameworks/transformers/5.4.0/test_result.png differ diff --git a/frameworks/transformers/5.5.1/Dockerfile b/frameworks/transformers/5.5.1/Dockerfile new file mode 100644 index 0000000000000000000000000000000000000000..e199d7ab03f3712f508f39badebc4689337f0c56 --- /dev/null +++ b/frameworks/transformers/5.5.1/Dockerfile @@ -0,0 +1,16 @@ +FROM opencloudos/opencloudos9-cuda-devel:12.8 + +LABEL maintainer="OpenCloudOS Community" +LABEL org.opencontainers.image.source="https://gitee.com/OpenCloudOS/ai-agent-container" +LABEL org.opencontainers.image.description="HuggingFace Transformers 5.5.1 (GPU) on OpenCloudOS 9" + +RUN dnf install -y python3-pip && dnf clean all && rm -rf /var/cache/yum/* + +RUN pip3 install --no-cache-dir uv \ + && uv pip install --no-cache-dir transformers[torch]==5.5.1 --system + +ENV NVIDIA_VISIBLE_DEVICES=all + +RUN echo $(date +"%Y-%m-%dT%H:%M:%S%z") > /opencloudos_build_date.txt + +CMD ["python3"] diff --git a/frameworks/transformers/5.5.1/README.md b/frameworks/transformers/5.5.1/README.md new file mode 100644 index 0000000000000000000000000000000000000000..3a036eb567100eec3163ed9e74335ac380a3cd20 --- /dev/null +++ b/frameworks/transformers/5.5.1/README.md @@ -0,0 +1,32 @@ +# Transformers on OpenCloudOS 9 + +## 基本信息 +- **框架版本**:5.5.1 +- **基础镜像**:opencloudos/opencloudos9-cuda-devel:12.8 +- **Python 版本**:3.x(系统默认) +- **CUDA 版本**:12.8 +- **核心组件**:HuggingFace Transformers 5.5.1 + PyTorch + +## 构建 + +```bash +docker build -t oc9-transformers:5.5.1 . +``` + +## 使用示例 + +```bash +# CPU 模式 +docker run --rm oc9-transformers:5.5.1 python3 -c "import transformers; print(transformers.__version__)" + +# GPU 模式 +docker run --rm --gpus all oc9-transformers:5.5.1 python3 -c " +import torch +print(f'CUDA available: {torch.cuda.is_available()}') +import transformers +print(f'Transformers: {transformers.__version__}') +" +``` + +## 已知问题 +- 无 diff --git a/frameworks/transformers/5.5.1/build.conf b/frameworks/transformers/5.5.1/build.conf new file mode 100644 index 0000000000000000000000000000000000000000..db399a5def7299852f798c7bbd062c468e19bd72 --- /dev/null +++ b/frameworks/transformers/5.5.1/build.conf @@ -0,0 +1,4 @@ +# transformers 5.5.1 on OpenCloudOS 9 (GPU) +IMAGE_NAME=oc9-transformers +IMAGE_TAG=5.5.1 +GPU_TEST=true diff --git a/frameworks/transformers/5.5.1/test.sh b/frameworks/transformers/5.5.1/test.sh new file mode 100755 index 0000000000000000000000000000000000000000..7f8912c900d7377f571507aa5da1f194f4b3c435 --- /dev/null +++ b/frameworks/transformers/5.5.1/test.sh @@ -0,0 +1,46 @@ +#!/bin/bash +set -e + +IMAGE="${1:?ERROR: 缺少镜像参数。用法: bash test.sh }" + +echo "=== Transformers 5.5.1 功能测试 ===" + +# 1. 验证框架导入及版本 +echo -n "检查 transformers import... " +docker run --rm "$IMAGE" python3 -c " +import transformers +assert transformers.__version__ == '5.5.1', f'版本不匹配: {transformers.__version__}' +print(transformers.__version__) +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 2. 验证核心功能 — AutoTokenizer +echo -n "检查 AutoTokenizer... " +docker run --rm "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +from transformers import AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +out = tok('Hello, world!', return_tensors='pt') +assert 'input_ids' in out, 'input_ids not found' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +# 3. 验证 CUDA 环境及 GPU 推理 — BertModel +echo -n "检查 CUDA + GPU 推理... " +docker run --rm --gpus all "$IMAGE" python3 -c " +import warnings +warnings.filterwarnings('ignore') +import torch +assert torch.cuda.is_available(), 'CUDA not available' +print(f'GPU: {torch.cuda.get_device_name(0)}') +from transformers import AutoModel, AutoTokenizer +tok = AutoTokenizer.from_pretrained('google-bert/bert-base-uncased') +model = AutoModel.from_pretrained('google-bert/bert-base-uncased').cuda() +inputs = tok('Hello', return_tensors='pt').to('cuda') +with torch.no_grad(): + out = model(**inputs) +assert out.last_hidden_state.shape[0] == 1, 'batch size mismatch' +print('ok') +" && echo "✓ 通过" || { echo "✗ 失败"; exit 1; } + +echo "=== 所有测试通过 ===" diff --git a/frameworks/transformers/5.5.1/test_result.png b/frameworks/transformers/5.5.1/test_result.png new file mode 100644 index 0000000000000000000000000000000000000000..f6435240e47d232060bbcfccde03c658cc5cc39b Binary files /dev/null and b/frameworks/transformers/5.5.1/test_result.png differ