update
This commit is contained in:
@@ -374,31 +374,33 @@ DS_BUILD_OPS=${DS_BUILD_OPS} pip install dist/deepspeed*.whl --force-reinstall
|
|||||||
# ./install.sh ${DEEPSPEED_INSTALL_FLAGS} --hostfile /job/hostfile # ./install.sh --allow_sudo --pip_sudo --no_clean --hostfile /path/to/your/hostfile
|
# ./install.sh ${DEEPSPEED_INSTALL_FLAGS} --hostfile /job/hostfile # ./install.sh --allow_sudo --pip_sudo --no_clean --hostfile /path/to/your/hostfile
|
||||||
EOT
|
EOT
|
||||||
|
|
||||||
# RUN <<EOT
|
# install transformers and flash-attn
|
||||||
# #!/bin/bash
|
RUN <<EOT
|
||||||
# source /opt/conda/etc/profile.d/conda.sh
|
#!/bin/bash
|
||||||
# conda activate ${CONDA_ENV_NAME}
|
source /opt/conda/etc/profile.d/conda.sh
|
||||||
# # install transformers
|
conda activate ${CONDA_ENV_NAME}
|
||||||
# git clone https://github.com/huggingface/transformers ${STAGE_DIR}/transformers
|
# install transformers
|
||||||
# cd ${STAGE_DIR}/transformers
|
git clone https://github.com/huggingface/transformers ${STAGE_DIR}/transformers
|
||||||
# python3 ./setup.py develop
|
cd ${STAGE_DIR}/transformers
|
||||||
# python3 -m pip install -U --no-cache-dir "pydantic<2"
|
python3 ./setup.py develop
|
||||||
# # install flash-attn
|
python3 -m pip install -U --no-cache-dir "pydantic<2"
|
||||||
# # pip install packaging -i https://pypi.org/simple/ --trusted-host pypi.org
|
# install flash-attn
|
||||||
# pip install flash-attn --no-build-isolation -i https://pypi.org/simple/ --trusted-host pypi.org
|
# pip install packaging -i https://pypi.org/simple/ --trusted-host pypi.org
|
||||||
# EOT
|
pip install flash-attn --no-build-isolation -i https://pypi.org/simple/ --trusted-host pypi.org
|
||||||
|
EOT
|
||||||
|
|
||||||
# RUN <<EOT
|
# other packages
|
||||||
# #!/bin/bash
|
RUN <<EOT
|
||||||
# source /opt/conda/etc/profile.d/conda.sh
|
#!/bin/bash
|
||||||
# conda activate ${CONDA_ENV_NAME}
|
source /opt/conda/etc/profile.d/conda.sh
|
||||||
# pip install optimum
|
conda activate ${CONDA_ENV_NAME}
|
||||||
# pip install peft tiktoken \
|
pip install optimum
|
||||||
# tqdm matplotlib seaborn numpy pandas scikit-learn diffusers \
|
pip install peft tiktoken \
|
||||||
# huggingface_hub spacy blobfile pycocotools \
|
tqdm matplotlib seaborn numpy pandas scikit-learn diffusers \
|
||||||
# xformers open_clip_torch \
|
huggingface_hub spacy blobfile pycocotools \
|
||||||
# zstandard -i https://pypi.org/simple/ --trusted-host pypi.org
|
xformers open_clip_torch \
|
||||||
# EOT
|
zstandard -i https://pypi.org/simple/ --trusted-host pypi.org
|
||||||
|
EOT
|
||||||
|
|
||||||
# add vscode server
|
# add vscode server
|
||||||
# RUN <<EOT
|
# RUN <<EOT
|
||||||
|
|||||||
@@ -32,7 +32,7 @@ services:
|
|||||||
USE_ROCM: 0
|
USE_ROCM: 0
|
||||||
USE_XPU: 0
|
USE_XPU: 0
|
||||||
CUDA: cu121
|
CUDA: cu121
|
||||||
CUDA_ARCH_LIST: "80;86;89;90" # for RTX 4090, all : "80;86;89;90"
|
CUDA_ARCH_LIST: "80;86;89;90;90a" # for RTX 4090, all : "80;86;89;90"
|
||||||
SETUPTOOLS_VERSION: "69.5.1"
|
SETUPTOOLS_VERSION: "69.5.1"
|
||||||
DCUTLASS_NVCC_ARCHS: "90a" # 90a for H100 GPU 89:GeForce RTX 4090
|
DCUTLASS_NVCC_ARCHS: "90a" # 90a for H100 GPU 89:GeForce RTX 4090
|
||||||
volumes:
|
volumes:
|
||||||
|
|||||||
Reference in New Issue
Block a user