diff --git a/finetune/README.md b/finetune/README.md index 430d618..50902b8 100644 --- a/finetune/README.md +++ b/finetune/README.md @@ -9,4 +9,53 @@ docker-compose -f docker-compose_pytorch2.3.yml build ```shell docker run -it --gpus all --name deepspeed_test --shm-size=1gb --rm hotwa/deepspeed:latest /bin/bash +``` + +## [查询GPU 架构 给变量赋值](https://blog.csdn.net/zong596568821xp/article/details/106411024) + +```shell +git clone https://github.com/NVIDIA-AI-IOT/deepstream_tlt_apps.git +cd deepstream_tlt_apps/TRT-OSS/x86 +nvcc deviceQuery.cpp -o deviceQuery +./deviceQuery +``` + +H100 输出 + +```shell +(base) root@node19:~/bgpt/deepstream_tlt_apps/TRT-OSS/x86# ./deviceQuery +Detected 8 CUDA Capable device(s) + +Device 0: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 1: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 2: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 3: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 4: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 5: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 6: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + +Device 7: "NVIDIA H100 80GB HBM3" + CUDA Driver Version / Runtime Version 12.4 / 10.1 + CUDA Capability Major/Minor version number: 9.0 + ``` \ No newline at end of file