diff --git a/docs/developers/hami-core-design.md b/docs/developers/hami-core-design.md index 7ce5c5f1..907dbe7a 100644 --- a/docs/developers/hami-core-design.md +++ b/docs/developers/hami-core-design.md @@ -4,7 +4,7 @@ title: HAMi-core design HAMi-core is a hook library designed for the CUDA environment. It functions as an in-container GPU resource controller and has been adopted by projects -like [HAMi](https://github.com/HAMi-project/HAMi) and [Volcano](https://github.com/volcano-sh/devices). +like [HAMi](https://github.com/Project-HAMi/HAMi) and [Volcano](https://github.com/volcano-sh/devices). ![HAMi-core architecture diagram showing GPU resource controller design](/img/docs/common/developers/hami-core-design/hami-arch.png) diff --git a/docs/developers/protocol.md b/docs/developers/protocol.md index 789e75b9..f575904e 100644 --- a/docs/developers/protocol.md +++ b/docs/developers/protocol.md @@ -29,7 +29,7 @@ hami.io/node-nvidia-register: GPU-00552014-5c87-89ac-b1a6-7b53aa24b0ec,10,32768, ``` -In this example, this node has two different AI devices, 2 Nvidia-V100 GPUs, and 2 Cambircon 370-X4 MLUs +In this example, this node has two different AI devices, 2 Nvidia-V100 GPUs, and 2 Cambricon 370-X4 MLUs A device node may become unavailable due to hardware or network failure. If a node hasn't registered in the last 5 minutes, the scheduler marks it as 'unavailable'. diff --git a/docs/developers/scheduling.md b/docs/developers/scheduling.md index b6ac744a..7e4e571f 100644 --- a/docs/developers/scheduling.md +++ b/docs/developers/scheduling.md @@ -4,7 +4,7 @@ title: Scheduler Policy ## Summary -Current in a cluster with many GPU nodes, nodes are not `binpack` or `spread` when making scheduling decisions, nor are GPU cards `binpack` or `spread` when using vGPU. +Currently, in a cluster with many GPU nodes, nodes are not `binpack` or `spread` when making scheduling decisions, nor are GPU cards `binpack` or `spread` when using vGPU. ## Proposal @@ -26,12 +26,12 @@ node binpack, use one node’s GPU card whenever possible, e.g.: - node2: GPU having 4 GPU device - request: - - pod1: User 1 GPU - - pod2: User 1 GPU + - pod1: Use 1 GPU + - pod2: Use 1 GPU - scheduler result: - - pod1: scheduler to node1 - - pod2: scheduler to node1 + - pod1: scheduled to node1 + - pod2: scheduled to node1 #### Story 2 @@ -42,12 +42,12 @@ node spread, use GPU cards from different nodes as much as possible, e.g.: - node2: GPU having 4 GPU device - request: - - pod1: User 1 GPU - - pod2: User 1 GPU + - pod1: Use 1 GPU + - pod2: Use 1 GPU - scheduler result: - - pod1: scheduler to node1 - - pod2: scheduler to node2 + - pod1: scheduled to node1 + - pod2: scheduled to node2 #### Story 3 @@ -57,12 +57,12 @@ GPU binpack, use the same GPU card as much as possible, e.g.: - node1: GPU having 4 GPU device, they are GPU1,GPU2,GPU3,GPU4 - request: - - pod1: User 1 GPU, gpucore is 20%, gpumem-percentage is 20% - - pod2: User 1 GPU, gpucore is 20%, gpumem-percentage is 20% + - pod1: Use 1 GPU, gpucore is 20%, gpumem-percentage is 20% + - pod2: Use 1 GPU, gpucore is 20%, gpumem-percentage is 20% - scheduler result: - - pod1: scheduler to node1, select GPU1 this device - - pod2: scheduler to node1, select GPU1 this device + - pod1: scheduled to node1, select GPU1 + - pod2: scheduled to node1, select GPU1 #### Story 4 @@ -72,12 +72,12 @@ GPU spread, use different GPU cards when possible, e.g.: - node1: GPU having 4 GPU device, they are GPU1,GPU2,GPU3,GPU4 - request: - - pod1: User 1 GPU, gpucore is 20%, gpumem-percentage is 20% - - pod2: User 1 GPU, gpucore is 20%, gpumem-percentage is 20% + - pod1: Use 1 GPU, gpucore is 20%, gpumem-percentage is 20% + - pod2: Use 1 GPU, gpucore is 20%, gpumem-percentage is 20% - scheduler result: - - pod1: scheduler to node1, select GPU1 this device - - pod2: scheduler to node1, select GPU2 this device + - pod1: scheduled to node1, select GPU1 + - pod2: scheduled to node1, select GPU2 ## Design Details diff --git a/docs/installation/how-to-use-volcano-vgpu.md b/docs/installation/how-to-use-volcano-vgpu.md index 9fe52783..bb5960e3 100644 --- a/docs/installation/how-to-use-volcano-vgpu.md +++ b/docs/installation/how-to-use-volcano-vgpu.md @@ -95,7 +95,7 @@ status: ### Running vGPU Jobs -vGPU can be requested by both set "volcano.sh/vgpu-number", "volcano.sh/vgpu-cores" and "volcano.sh/vgpu-memory" in resources.limits. +vGPU can be requested by setting `volcano.sh/vgpu-number`, `volcano.sh/vgpu-cores` and `volcano.sh/vgpu-memory` in `resources.limits`. ```shell cat <